var/home/core/zuul-output/0000755000175000017500000000000015146443703014534 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015146447554015510 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000235037115146447402020270 0ustar corecoreOikubelet.log_o[;r)Br'o-n(!9t%Cs7}g/غIs,r.k9GfB~"mv?_eGbuuțx{w7ݭ7֫B% oo/q3m^]/o?8.7oW}ʋghewx/mX,ojŻ ^Tb3b#׳:}=p7뼝ca㑔`e0I1Q!&ѱ[/o^{W-{t3_U|6 x)K#/5ΌR"ggóisR)N %emOQ/Ϋ[oa0vs68/Jʢ ܚʂ9ss3+aô٥J}{37FEbп3 FKX1QRQlrTvb)E,s)Wɀ;$#LcdHMeBmFR5]!PI6f٘"y/(":[#;`1}+7 s'ϨF&%8'# $9b"r>B)GF%\bi/ Ff/Bp 4YH~BŊ6EZ|^߸3%L[EC 7gg/碓@e=Vn)h\\lwCzDiQJxTsL] ,=M`nͷ~Vܯ5n|X&pNz7l9HGAr Mme)M,O!Xa~YB ɻ!@J$ty#&i 5ܘ=ЂK]IIɻ]rwbXh)g''H_`!GKF5/O]Zڢ>:O񨡺ePӋ&56zGnL!?lJJYq=Wo/"IyQ4\:y|6h6dQX0>HTG5QOuxMe 1׶/5άRIo>a~W;D=;y|AAY'"葋_d$Ə{(he NSfX1982TH#D֪v3l"<, { Tms'oI&'Adp]{1DL^5"Ϧޙ`F}W5XDV7V5EE9esYYfiMOV i/ f>3VQ 7,oTW⇊AqO:rƭĘ DuZ^ To3dEN/} fI+?|Uz5SUZa{P,97óI,Q{eNFV+(hʺb ״ʻʞX6ýcsT z`q 0C?41- _n^ylSO2|'W'BOTLl-9Ja [$3BV2DC4l!TO C*Mrii1f5 JA *#jv߿Imy%u LOL8c3ilLJ!Ip,2(( *%KGj   %*e5-wFp"a~fzqu6tY,d,`!qIv꜒"T[1!I!NwL}\|}.b3oXR\(L _nJB/_xY.# ſԸv}9U}'/o uSH<:˷tGLS0l/LKcQ.os2% t)Eh~2p cL1%'4-1a_`[Zz㧦|k˭c ĚOρ_} Ewt3th?tvͪ{~;J0= |JUԍ;Iw}/9nh7l%>'ct Հ}a>-:(QxPyA Z UcÖgڌ:8cΗ|U1,-N9 dI [@3YN%:ò6PT:”QVay 77ĐrX(K&Y5+$wL#ɽ 4d-bbdAJ?w:P>n^2] e}gjFX@&avF묇cTy^}m .Ŏ7Uֻ󂊹P-\!3^.Y9[XԦo Έ')Ji.VՕH4~)(kKC&wfm#Y~!%rpWMEWMjbn(ek~iQ)à/2,?O Y]Q4`Iz_*2coT'ƟlQ.Ff!bpRw@\6"r+i37Z_j*YLfnYJ~Z~okJX ?A?gU3U;,ד1t7lJ#wՆ;I|p"+I4ˬZcն a.1wXhxDI:;.^m9W_c.4z+ϟMn?!ԫ5H&=JkܓhkB\LQ"<LxeLo4l_m24^3.{oɼʪ~75/nQ?s d|pxu\uw?=QR -Mݞίk@Pc n1æ*m$=4Dbs+J \EƄզ}@۶(ߐ/ۼ𹫘qݎt7Ym݃|M$ 6.x5 TMXbXj-P\jА޴y$j`ROA"EkuS#q * CƂ lu" yo6"3껝I~flQ~NCBX`]ڦÞhkXO _-Qy2$?T3ͤEZ긊mۘ$XD.bͮW`AީClСw5/lbl[N*t*@56."D/< {Dۥ sLxZn$N(lYiV =?_e^0)?]{ @| 6+#gPX>Bk2_@L `CZ?z3~ }[ tŪ)۲-9ֆP}b&x Uhm._O 4m6^^osVЦ+*@5Fˢg'!>$]0 5_glg}릅h:@61Xv` 5DFnx ˭jCtu,R|ۯG8`&ו:ݓ3<:~iXN9`2ŦzhѤ^ MW`c?&d.'[\]}7A[?~R6*.9t,綨 3 6DFe^u; +֡X< paan}7ftJ^%0\?mg5k][ip4@]p6Uu|܀|Kx6خQU2KTǺ.ȕPQVzWuk{n#NWj8+\[ ?yiI~fs[:.۽ '5nWppH? 8>X+m7_Z`V j[ s3nϏT=1:T <= pDCm3-b _F(/f<8sl, 0۬Z"X.~b٦G3TE.֣eմi<~ik[m9뀥!cNIl8y$~\T B "2j*ҕ;ێIs ɛqQQKY`\ +\0(FęRQ hN œ@n|Vo|6 8~J[,o%l%!%tyNO}}=ʬ-'vlQ]m"ifӠ1˟ud9)˔~BѤ]һS8]uBi( Ql{]UcLxٻa,2r(#'CDd2݄kTxn@v7^58þ Ţ&Va%ĉUHSR0=>u)oQCC;^u'}8H0]+ES,n?UU{ x~ʓOy_>?/>l8MrHID2VSsMX^"NۯDc558c&'K0L /C5YDqNe~ض˸nErc֋@aw*r܀0 a {RQXV-/p:MP\<=<^越a/bz?ܓvjIg3MN4:]U]STa,@OKd9A7ޕ6ql?N/e1N2i)i_TA|S2G4miBȨHM(2hys|F 94 DNlϒòκ-q|xC ,gKDzHR%t+E/wd#礱ºȄWEz o\JξB.wLKZ39(M +(PWՇfR6#ю3Ȋt ݪbh]MTw䀩S]'qf&)-_G;"1qz퇛0,#yiq$ՁɄ)KٮޓJ|̖D?:3mhW=rOf'/wѹ8BS8]`;=?,ڼ"ϴq*(A7? /W= #^ub"6q f+=^OI@߱^F[n4A#bYѤwd)J^Z{*ǥzw73LuaVad=$6)iI gC~.1%YmҪ+2gSt!8iIۛ*JgE7LGoş\bC}O i ycK1YhO6 /g:KT sPv6l+uN|!"VS^΄t*3b\N7dYܞLcn3rnNd8"is"1- ޑܧd[]~:'#;N(NknfV('I rcj2J1G<5 Nj̒Qh]ꍾZBn&Un' CyUM0nCj.&Oڣg\q0^Ϻ%4i" ZZG>Xr'XKc$2iσֹH<6N8HSg>uMik{Fm(W F@@{W+ߑ?X2hS4-=^YgpUHެbZ!y!ul@ڼ63" ۩:6=TZõ$E,ϓRV|G&$rr;J TtIHFE=RȬ]P pLm|?$%>Eü%mWO[>Xmw,*9.[G n >X8Ī;xW%dT:`ٓ~:VD)O>UD;;MY,2ڨi"R"*R2s@AK/u5,b#u>cY^*xkJ7C~pۊ ~;ɰ@ՙ.rT?m0:;}d8ۈ ݨW>.[Vhi̒;̥_9$W!p.zu~9x۾vC;kN?WƟ+fx3SuKQqxST Ζ2%?T74a{N8;lr`$pZds=3jwlL Eڲ t|*n8[#yN SrA GYb8ZIaʼn8 #fg3i`F#5N 3q_M]j 8E!@1vցP7!|+R@;HspSI]ڻCZUcg5pDcIϹ,oN-_XI,3\j ]ٟ5~' SuipA!C厐$&k7dmhz/#"݃,YqCL$ڲ`"MUbeT>Xuv~4Le͢ }UVM)[A`b}mcE]LCEg=2ȴcmZ?E*-8nhױ1xR2ϫCya` A y!?h!9yL%VLU2gr26A!4vbSG ]ꧧWp/ &ee *w$-`J\ ptǣC^p#_`{ К8EW>*(D{ٛ,[fnY𱹞M=6&$<,"lX-Ǐ_whaE 98 (oѢ/Р΅ 7ցl6618ł_1/=fu).s¯?.S[{'g=Ҥ):d8h\y6]t1T7IUV:;.1& ,5΀j:<< +Y?58In'bXIǣO{&V\DŽ0,9f O_"[l:h¢8wݓ19\:f6:+ .3}=uvKc ٹeS<>ij(o'ciS<{1$E[nP b?8E'xv[K+E{,Qƙ1*dcs_Z'407|qBOgYU|U--sG8`u! qGYܷw;ȌCPc_|(RaIBKb+{P.T! =ĦiTob d<>SHr][KqWs7ѝBYǭ~RR"p9dFg|K- obY_vM 4>/]e/dy,8!xŋ5 R<^mYo 3c9(F?he:9[v~\:HP 8'k0t1A!jlX)v/L+NhBUx~Ga>Z"Q8_jTLRKtL L+BT-҂=ll魳Cf[L胍̎`7rIkzS- (J[(6 b Fڨ?z ZvƂcUkdύ-׫E7e0ϕw6K!x^>$ N7_HNe"z`0A*0)QsUN8tl^N+mXU-q2EDöbVm')fpOj4r@mί%Y!/7$&5n8TK&cB/5q0kS换ohN\U:=ƹb ZFng&# pil;I-S)`н' X'fe~9ƟE q"hz7z^>%QdE6c>!Ƶ *0W4Qޣ@>lWN"A X5G-nm.8B.OI[31,j2C_:+Z|p8!81,w:$TiDz7Զ]co~@>\X=8OZS׹U>8bK0&V\ t!ku`k\c0h&)IV })p| +fjI`bv0ο0ߒ"G̾H`yY/'O"Q_4IAٍKK7'l[ QVm0c<%UEhZ_.1b~|n2ͦ_DQP/2 re%_cRw~r9_7*rn |c/V&, ,B82^WK9EHLPm))2.9ȱ  QAcBC `!1X.")+T4a|ry !w5yY~ Z3;Y[Jȧ q:iǞ/&T+uIrd31V_Uqb0t/ %I[hq ҕx  O U^>wY~ -`%Űb`X{S38W!`znOF7/.C!Pu&Jm l ?Q>}O+D7 P=8! ЛN_[d0Yݎ@2!vZ{Ibi/^cygwpГzY'Ź$[fr;)ٖf ՠ5Kcxg* E Qu{$Sڸ7pOqz|T.wK?AW"ϛ# {]ҙ%rXgs+'n*PuLٱתk}Eflѓnסn}< a+czb@U#=h7) ^EUB68!˂co\Zً}q| kܸ98V#%æ~c ՟Sxa'r?(3{ xRP8_S( $?uOk|mbP\vە晋cLz6r~CMp!l?o AMҪ1Erw}xxݸǻ*Wu{}M?\GSߋ2ꮺ5w"7U0)lwۨB0ח*zW߬V}Z۫ܨJ<]B=\>V7¯8n8gqOg-?[~,;n9 |q?;nmCgNd_@DvzΙkkApȏ$-X )@Wbuw;YɊ|ZԘϘee|Bf".4IK1`b $1r Qz?ۥ1Zm'G+qYcYl YhD$kt_TI ^pM0[||V O8Z-mYO!t{O`% <fVlc/ڀs-Y:dE$E 4ehMÖ6%%<@fpKQ31pz}(6>C,HI~'.Op% 8$ c*Dp*Cj|>z G` |]e*&q!tW qBAgPS}E`ǧ́JE_!]8$ 5NxK5?W~>egI+ I.*F~H!Gf"LD&Um/6Gd#fR*cz,—rw=3~rmsL&ԙyW$)bG(]71wt!wVf;A&Xi~!qvGkJ+u7Τ!=ljK1SpHR!:YbNqǪP@o`co <عۀGvm۵8[,)4\\=־?Cn._) cxF;Gu'n [&8NJPjeyY!UPۮu2Ja>0e0Vwގj$]p^翍+|Z}B9{bOe@7'XL:OEUn|23 ZC]pj|4iBN _7C$h3`|Ulk\)WV`JBr^J |rt~Uv1=UF6"tsiT$U>LNÔkXrFscC0 X%E+o*tNU*׫t[=EAV$=q[mhm"ro5jq$j!;V0eOޝ4ccc2JWN>7q;aEI"*o/!8.Ԧg]kYkDΡ1R:Vݐ/CBc˾[ sX 5)X<sefS ;W{@dl)t%عF&4D&u.Jm9c$A$ Dfj-أ^&#HȯTgرBIx[ 5)l>MR2ǂw J0A*1eJ0Gk@5^VtY"Fbi[pdNѱ X^E 2侫}>Ir~Z"+A<Lй nR)L-w}ln2&0Ǚ'0>Ƶ-T4G7! H'ͺ|@lX)+{{LT/sekϙJ$S`hl nOcE(Tcr!:8UL | CtQ7jknnmK0ܝ0GGL'Pz7e&Ӄ)^ʮLW's%iJ `uv6%OE-56 0-v/Xŷ%r׽nl-ߑst2S%tTڪ?>>{2])|Ը>U;Mѹ .Vfz0Ïd0>7. ]|>TT%69dp-*VVK=$l&~g۷&i"&I2Wwߎ|7.sW\zk﯊溺^TW^T\*eqr/^T77WNZ7F_}-򲺺VWQ77V\_v?9?"Th l&ەQjaXMʂczOAo.$@bѠLtT=CKGX ^/) 5 'CjN7[<Uq/ҿO6$dxs` Pj=̀*Eg.6_~J/8V ?çMR3V:<}xXX]^deU/Ӽ8Ģ8{Ʉvb/{ Ih~|Y4sopp|v1f2췓$?mc+_B73㙛8}xNTt3]U˶}9o]|9ͅyߵjEM`'y:x'3wx E&{]9vu8B7ws ?)_NEa$~S)C;J!>%m!B!>..bi9\׸ ξFawG"leiֽ:Z]oRh!BWo B7. Bw.t´< 쵣ͣm: zԂV#7"3`Gp u\!B Oں{}z8~I7S7\۵"L$?T込F2CcLšB>B%]x)2~wTG;"%=[q7u;%qTϜ=3 {7e,ŸjŒfg*,V ^/Vyxy8b;`h;tZQ]TY~g,\y--2#LXIaiZ#@{X煼lD6i%(oJMfcP%(2YS?˸)(V9ye^I\@aMBu1~Te:K9YuQ]լe5kW [u^{Hu6rZL [-p2}tVڤ[RDM%{%'9,y)&f* \L'I,e@L|5uW%;h㝥iZW1=\P8g ~Aɷ,'i ^g<2HIGh g O~;}>#A |Ƽ2?tW_]28AFYdЈ Z<ݓiףwMVy+Ga8:k:tS&)aSG5"߳(d6y.2lʸ}UrNN[-ۣMoOO,̓YՒ9Ou(/3&9 kQLM־~hj=o8ċe]ѽw (zʻqxt*f8A$91Xf/0+u)yz3I`9 _s,<аogZEe: X/x*;z#G-퉆`BNAܓOډLǥ,^=,ӵ P 憦/Asq|Sޛˏ;L5u^m&8SX:AwNeXH`M=͍sZ[Xa骍Ѡ+yeRПeN][r qa xs wx#q,BH`l>;!ap *Fa>e$3ZƧLtBrvb@ GG|￲oyP[pACA;.ʦ7b$a1ZTvӻ}Ņ$*}^}:O| %0۔(ޚwgwHm+α*MȲ bڍ3ꂤyS3^T˜]6Px0 ;2}짔YAN*yR/O2&a)>?QRCGL.Y\.CsW+ ) uF)HP,l~J~$k!zp-՛+yɪT}Ew[S@M8>H&VѰ*=iByźjUd; zi=1?ēŊ#zkmm_ _Tȁc5_^O4 >m="Q.zZW$N4Yÿ )]/e]`GO*<@k` e'R8IM\шo<(l(l&ǤqpXGX%u'g"OgXРJnOojí"g44bڮ".g.cm_E{pMeD_g;-ahK'Ui[jWma)~Ө**LD]*5W)DFmoIb9ΎX{c?) Aguh}Y鐋U\m+ a ȥ}NMk{ϥ[YLL~ˁgaWm -[2ʪ!ȥ7hw0DϬho+sW1]<F7CƘ ǹhDiE#l@8jG67Gp8$k>ŊhieN2]8Q*ODfj5Æ&$19N (mɱOsY Ae;*a?o=gO߉lwt>=_E0Pj(d [65fruД=^zzTR.B<GxDݫa V4,oGbME {ZYweuuʵHHpSJ.yV{=3TG;};<{3[fe_ ր0-omԕ6l[i-^dy)7R ctBcxFPx1U߁bu eKލMSa))i&P72C Surzi˼䥔zv1Њ'P՝k_}6ki\ 2 lKт89<0mшikx -Wi}KE0)xP-nmxpB,#1R,f>D ݺ.hGxlF 7Mye"Z]/Bm/Zv'tx@2I~u.ͩ=߆iTU.}2 p&7]$"b;pP'RK7I)tm Vn%l}W\CmURnJF0[GC&4MCMs%&1lAMԺ$*g[a>D Wy?dwڶ}%VK`ZSC+[e-l4ri2Hz /KE w/pTɎ #Fqvky]ϳPgۨ{\W@AvVY&o}"hhl%K RJkj B!2"LXʳ8U5eBbʊm³",9GRDCx6ΖXW Ј;0 h\]MF X8Phay{n˶6/A4bs3D^|3@b|)C q)k3]Tb N5)eTƛ 5Y.ak}Sf%,Dyؒ u/8mx*Gp΍ OWCEDkX;Wo?&M,FOB4bRӢF}ޑXSɹjdզD_yo<|wFW^].ro誀^Ovl< Tqu&>놱L+8ʾTKрzk Wț~047*yZAzA1vahd{*}棻ߞYܾ,"8>Z(H#_9m’4*۵?FA9h-}5T孫4@qLlXw͵lGf8Mezl?>#m7V@=5}vֳ/`sa P0a- a]o0:p}PCDZ#*0 ܾ;ཱྀ8.0LC7\ \[ϗ;Ђ ;CPuhF{aB7|YaJb0PpX;F `cuaj $2:oؑ{#EG"xZ ]''4[W#-*0F')̥GL 2D\l_Mfg*X^TYI8I)jwz$*(٬@]u4?Qs4ݞ).mR;Zt`ОG{&ѭUx]w+x,yewXO>>'9k`~2d`~tZ` t3-ߵAuG!uC߀ߩw=`:t"b(ނ!ηM}By9 ]taEl//k|B0a7ܔ AA/Sybu=6a4E( ^4Ъ[1B/{ ەp;nK"a]2! Ofb`jXBCЍX6":b,vJKds/^Vt-L0f`}mu t]CdE[uP)7|as``hǯ1C; FF&H?%D"9hx~J!k*?(t% q"wFRj!|yj#Z<; P8\qp b#d~;os9- {a{Y $byrxt'P|m96Áϻ,L|k- 퉐u#5.IbQX{jΛ$bt12F8qcn!)7`o&z&Y=͒J{I9{xF~ѤMY6+IT{;f/T_nTk$KhKHO&lp"B)6-Gn'u=M`vcTQ/>)z_熮!ҁ2OB(}6!>5l)\5 8:F}Z5E9dO't6`6+2TGΟdAjK(1͸[75dw3 HyZaM+wZ։aۈ}>ASWkgM:KYh:+ϻ ]Ao'ٸ(ѫ=eyϨ0$Rxє^gדɰwb*NsfoN[ K@Qi~4+CCdk -@E.4 B' ?~6 ,!r#-0"Wqx6.g%,v@ )alnL90ȱX$6 Ԏa`,;yf`}Ũ'jVX8#$o`OPqG~^=Mw+kaKo޲7^(0Ao:\.p7tXL{tؖnز G}T0 [g+[v}WQVCP|Ko>Nl;qY.E(mBxAyֻ܃k}Fޗx*oWME)YqZŔ{݈'Y:5lqvp5j?LSTGyY4 aXh-k6(*z*#l+uٲ..AP5X5#*"ms6pg%^wDZnM=<RWݍE(JĖڳk;ԋb%C(TLlJN ฏUqSٵՏhF $J‡>Ŵ=̶tqꭜVꈮ;x tc7.t$d7(2hm-ꊺ(\\^"˹b (` C8w6U4te < Jw N(ݏPDB. )eL9Jwid2wuWuw ݝVw?R'R@Jz=PoB*lBeO$@J?PBUB v'4؏;@h;~O$4܁PJ(߁P;|?B ;*V ;*v'TGx"bKq/JXW x@_&HB(Vw|wF]DԺIq9Xj`82qWE=|v]g$*؇w ;F$&qTXfr-@1geBCҿן3j{awzn/Kֻe0}瓬P5`xg 0-*#|8'`=4(;.u?vkx(V} ɨ"n|U%u "iY$},T-\@,-L"U2/En㇞0 Aci1F͐G2DU#YT&DCDvjNgwx߆9DZz$_",A ~YDYM{YzeqG~W^5ʓ*5xK(ӸZ>+#RR`sKj˚~yQJQc\tE_jn9:q6pIezYҢyͭ h럯7aQ&C틇;E*ڭMzwX!Ρ/q_6Yl~Y<9X*b'u1jpVz]jb/YU}'x4{615\ǁߕqüWI>9˒QV!PL[GK]23FO 2aoߚ _Y`!՛rȔ4A@o>('2L&l;+-OcW:` ֻ5HG 64 QiK>ag!XlM>^lLdzz_j?% Ys3ͱ`Vw{VQ\Sdq8H<*RhA[]tLL`3!g`+~4]@t#_sSSx+++=_JdR\DVRv ўo^y{W,Ӑ>i-<~]ʡ&tzBg@kGZ @ah*RM>A謚h:8c ?f0Sm&5 ׂZ$mVk -F`֯{vۢU 3`E`(c A Cbٴ0C%2Ft]T3[JFXL +30Z1*KJ ZJ ?c8iq;PQ 3,Sxj;e|. 4ge$浸P32?A8ilqrQŝ&2lU7զ)Ty73Oߧ <=z3`W\,'p@ѶnC,T"'ArR)#nKUN:'\۪~ms+eC^f}JEvQ|N6Ie`EKG*gZ:d0/+46ʵ\O ou9ϕAˆpt?B~%sHR2.ϡmhk/r0@~Xrf0LS˪yʑ3Ue^*B1k(g:T]*1`n'@jo$7I\e<ˤҀ\k\/SH5-ǟ+%ݻf)l%T<8Q7V%k_'Fpi䝾Xy`']PPʾ2걤R&(ᦾ<5Li%?;S{Nي‚f-9h+2'rЎ"ʻɖՙbraŠ}nс1Z\u#~LPRcΎM ֡Ҧz7.j͖yԺ5UH5O0p>af8T<AnT˳VJbunFڣE"\FjZha52H=gd1,*W6&2Z%ǂMda5bv؆R>kjἍyÞ8oo߸ts߶s7:kPvsYC4%E1mc7jڶ->j~ Q[f&Bh.,L11p|;A,l0"R^Ex@ XmZn[8.`ֲ mI=L90XJwH,C)o03Ɔ9s|{YQתu]*ۺ=Ux25F 1x.l6H\c'h'$ ;nff;R֑vLA˜6:G$g૖ɻW| `l#ٍ~ 3pHf1S}I ({qG0({uT\*,:ϊ()DA|;K6 _`7(]+YW,ͧR"^k&}4[%r$~ P8JJmWmf<4ԡC1gڋKd_ YWsQRdBJ81H^<%,5Oŏ: !RqI:y~KH-g4}DA5 ;-gepb,]tଛ=z}I6IѾ;Xkˑ'<=l J$,G x 6^v1CӁU8߅?i"}vtTREURSco?lH~hiE]v@ @ꤌ@$'땔Yc쳔G&Pg)8c"P QJk;1NmyMpT}g//XL3yuN FƙtPP2͞ Ns`D40<`̐n˻vd$dmu/Y}pVEW%> x.I[N.U-Qy5 AYAVLH - larf9De 6&A5T}9Q87tM o(Xx`væ \1cۺ?~jH웄0bVd Dz2a;P98 1Hw=ngTBv+7 N:qDǗ8mcK*H_BK&5/ @Xgf9dvZ3 1*S RHuR)̑9]<6s.ہ=F] k͛BI=v(v}kFz5k Z ̐կ/XW&)=C]!**\${!pyE !ZyUtr/G4T5ǭ2%JxȈ/Қ\J"V# I;QqC %s I qM[3ڏl V{)3s"5D  LЅb !RC&-eL 8Z)!EKJ J5daiٹRF0MpXv.Q_k<#Caz/ 4'c-ί e{8G'B }fH{˥ a̴WVoyKR) ()k-%Z!3w}= KkR:FW2l?C1VCݹ xL2]͟(hZ$Ϙz䃎پ rtReGb<|IlR<4!MJ#P%'+w&9=s #liN'Uo8."lk@3 dJ+|\E:FX:$^Q xZJMzRA@Tb!6AIʣK/ߌY.04>4 ![(Zo!ENNXBlB {kq6kt-Պc*0~6B~;WM Q2yj{iOp4N 'k=3r  է&Z1NN[a߿U[{W_UbVhe)- ,.y'EON@%aǪUY"ן֔65Qn_ <+l,^J3xd4}*gO> :R~F|rsN*]bibiEɚ9ZVXF6Qd.c gRv({99){҂'=}Kn]ٴʇGC0w{'y$9/JU#AG %&#KG0 ->t z\zr4?,dKtnV;s?hI>dRIh{RZT|/9%2|jc R6IEcPò 0**\${Xj@5֑4`̴޼Io7/>;buCܴǗ$EAViF<.@|b0M|Kw-U]!y^y q׶LbF@6`,־=nMp /ףb ~ꝱ o/YY5~ieoκza(ei{6m9~45aUoUeeV/\5}M++A lUguΰi%|gM,߹0]zBIYA*ţ!15դ(`D  Rۄ6Z1\iMc[禕x דệU GT ɽ6>p阡bIٵtuT͡ SwR߱u×RbF)Ŭ?/M}x_63REoBﻩsI !KXJRN-ODm9w`4 yvk&Tg2s#U u-ZW ٧\B{dnH)-nffPY?lgٶMŇ}[-ԝUF(I.XOj8Ĭ7RƐ}ƋNI]j1#/YXޙ&8jۍ(pd55"^k&ȯergfbgNOlC^Ծ\tp O4ѳ'6+KRQcTq`3*U_=ex@'Ru"f"a̐_g$jΤF>ء^-z)8U{9omm56#Ī +ͅt:h"=ޗ&8v-TOe$TiHkiJfP%E͎P&KuL*=ඁq U~u꼫*FW;/~3h}8ѭA7ZnMDmmG=9.`P~|M @q6hi{0_runNi[qYuu;a@ß;>7%ǟ"w8 VL6hiEގ/5j*4'w}Tۚۤ&޵FR:ur4վ)FO5 nmS^e2}׀&G0J%-=pTD\oOrcl3_'7l,do&,=p܂Zkfg%zdָ#N~]@QX3{sGU lzsq'Y Ysv/ 0ʦfϠ@?q+f(1=BARLEQ8>c) ]Ĝ1/$@TuYz6J7dW(d2d<䚵H)D;ZY4sʭz8M/*2 ӊ)htA>U_ g8_{=Sǵ*\ :Gٟ.=58 |v\voހ&ύOIN~x}iN/@tରluyr2 -ҡ*24a6=u'a0ɛB4H"uqik|c9wC{w1^ha~r 0jk|Xhc;>v]f#WdzA $ګȁXA4F0»_؀8Fʅ] Dϣ }y@v%k`?w~6S{?6-=As4D$rU)*J+=zv6D0d|Y og4Zibx%&\֮,ޱ95qg~8R_#0s֛xgDeW,fn翗^q ?thjʧu9W"W7gz QlwR7@Wo$r_(H̥ѝ&:ukd2.hrKw 6b3˹9+-Z~\Yɕ^pmgߴ"""O -ϳwge;gЁp5vq"HsUR?K3 Y( ϕ\d./Me畈X|Q;])zdڀ^R;/!_tsi X@۪5ܞebaMX sX,}s+,?=!+s`R/~kr6D*JäH3P4Л',$0K/ӳq1L qeK8ѭ /N췱h<a-*QlvX[4`[#٫٫9Rŝ p֖ϒ)%Hr#ʴ2! Log?ͻ;`PPŢSqi1zKy{V;. QO ej/6BTnEGcZ/o+X:ny{1!?Ņi)(p猔(B{2[+a1]Av7-Qf@mx,7@w t`) +ý:,!pM.,F U=S&L]YrI s/E\*N<?X6OɊ]E$k3'+pJ^/Jw;PKe )^U}Bޡ2 K <\KRPXm3s٦mZ]5n+3s_YY;|V]4{ҭR_p:X_aDu(+:jr=B4ūhPr㵿OS؄"DVγYY&<滓[D "i -LD?1-6ߗoф%@ B /= pc[\3f-jK3ypG1ۚ+R_vN  [=M-MF8,^m C܏J?+h<'?<׻7{7k.@yǬ9|fGymf`~n@t/+q 3tk5p3@r6!Xƚb{y}1ۊƍڎW Uh[6Tm g2K48s '0 V#ºqqڡ#]_\Gxspul)U>͕0] w0~:ԏ~mUb%@48%VF;)N8!0mXD9ÞB彐5(oMB6rZ=^cJ?aހ!kZY{i!p3׺st8cZ猐#9 cQʑj,417E$T rט3vԊ3vtp=إL #_j3n%u#ISxHqxSoȇEeTy 6KZXbh9C Ek>7gJ<-XV#kxzJ&c)s% jH>Ymne쨝Ѳ'EX}vhs &U+V48E%eՆv zz)C\ ,;4R? " #:28f lmo9X9GsJ,v&qffQQrg1}TI~\Ud2dy#',A|K~7V:oLǸyR#jvqK^pvYyX~,L;ɳeJݧQ$W@dŜ+5 We[\ì!d)B_JVE-u3?W'?JV>{pvx=;X_}Ӌ=<4TVR7fj_-͡NlcUK Z "|Jt@H3 a *Exؒ^HO%lFu+4?[́,B*z6P aN[3{+Rxe2`@] wkuTkh;ېKb.ϏD17z ~54 C;7K.߅i|Rƨ&Y_{"dQLoMT|^]WZ[\_y{`p7s?G~ߤ`JܹԸz<>D< Yf{Fψ,ۃz1~]딂BsXL?yΔ´>Py\C|>_g%&ɳ.24X`0DU-58܋uUVlj>/Ęe}#LjM5n[„x2[jha4& ~8>ZG"텎:5>%1:ō+!R!BS8k߽洣#m_-PkYL{vsy-1{bybdR[^[|^R^OP0^p=gDph y)UsD, |0|z䳯Y SiݩXn=kA;$AuF;0F \Si- Iua-g[0_ǔX;!yG5VwqÒlk;mdg K2:%%C"X2I#Z-nt8,%/T# 6 5NC k0`O= A1#BpZF@$ĴW"<A} ;}^4/#L7%9KA'j-nܜ T;s8td=V{}^Rq}ϋ*lZ:3_^>!})+ևfKV#.Nxɔ)c,Oa&-ǘèVL7x*̢ 'f* KA(UOa&cDq[e]6hy}#D&`c&]fTI&<If.q0#jC $sum`J69T ASzByيe1rd~+0zo.NxbҰRk< Ѱ`mAq"S5,n (!͕7hAA . (B7a0X$ITݪ!jwAvx_Z 0 i Hqjpá*BbAV'JL {(qiJt2LfH Ij˘P߾1pNAC$ KhڱNd 3YGbOaIۗTY|!儡,M&ib0[W)U%Ÿg3-ES$ %i,E,:v@%5A@V;Zl! wZ2=,nZ D9RD [y-hvΣEܹҊFQ2:X ^M7[`EXN b:5c}4Cfh +r9-H &}O`R`2~Kz-~vhi;- 5$nى(RZRkZ<ѥayuNT k*OtNՃj+{ oS_f0;%\odciPi0aZ+PV}B2@!gDHpe[Ud( QYvt(V1c\jC!d CAL:]H/ =GkݰjVkRc9>7TyXВבaY)FǘaQcrPRN4QֿD_EN2*Wgg \צڽ#q_jZG?>kTܥj_`C\biUb}ZoYω-򞗲BPAZ;8G? ZMC1.C`Zeݣ ] LVĈ\_g fn]}|4?IvG׽Kws?G\dLV`{nXڂvh/W&,OU RGRW&|C`|>K/kGq'THm*ljy^ 5D1bN*M/z C7N¢,R篧`F]}"7~W :gOc*IP2γ8.c,PkICT Fo7}myrX߸r5_V.W0>VRچwU-O>g<%c-g^sA K09ƚZ}6{LPvī*[ xl;|f,L}Yنhlii QS)s4KT*Y2.VbI V)}+ֈamn|!N"5%ZD쭓f5K%YOgV ֮RvK0]ݍ++#؏~^; U:F߳o˫D 36cv2o:{ O+\6w2xCOsI|5߬)(+hMI>[ř+If|aꂖbZСqjxf᚜0?.MQ6@H cW$a <>3AYdڜ8#?]Fs_.̋-f߼$+OfcӍ?zet*U zbwN ≩- ҄[5f"O:s4 lF[sV{7"C8Wl{0(^B^ں ^YZ 0e^x. %~j[eI !"=-w J*&kk,W{عSO_[oή-;=ؗbZ n q hB)\_eJMͽ>1޽62ʳ=hun=4ݕ}vv)for{'.|^N)mc㊃;5r~m8LD\2sR{۪oοlI%&;N=~\_&ŋQT>VߐbeZN5c;C[Euw8q .1j<݂kܲ#-^λ=?D%GT\G~EаOC;)GW TS & тnw%b:JFCTk&~ulu޼菏?npe5\pqS*f 8qqw}>UW-K?? K ;R!ܵFYkX4p>m]t׸=_z`A[fBc$:I̥}Wӽp!@)bgÅ,O bpb U *g4q+M)KmL yu$*|&5Xg>ЪյM5pSyq./V4_mp{4(P1m xZ GAoV ͡"akd 6r/8a⇟xMn.w;}]->j;׈_k x/m#׼-1;$e(Mvb^,VUQ.ܛŬ~M]d߂Ť 6M9Y.]mAMn6)Q1Yv? h$3hɇ7L..̼ZnҧfV.9/VxVp'ŷ}ecV6ML-ҏDZ2\~dGN7` DeX^cHsj}LJU= /7Co_*|=Ð#̓Gٯ+ jLu̜O0U0b* @iE1\[ e<4dM ^6IԭSwbSL-|DX>ܭ杜RLKbLP'uF"~?(.EK}%j<"9VF=۹BrMޥ4-W1LFȗt3oߟ;8?K6ψ4|` c;NFL_q<7GWGI&/#Hn 颐,1uD `.̴%FȯE~V/G PPM T5'%BjxEZQ%W|1ώő/טn@qRC&`(^%ꐜ+TRPGM? ŀ/`ᕳ%IQ|~ITi'ӊ'2-c@f:|!t )^H@ {S !ZЉ;XG||wyq['op"ނeZ(JAEb6|!L 3Ri4DI"i -1(VpKm_64p#jKqZeIyeRғ) ?N0Kwt](Ӓ)9mͣ(Z /IO}oL? RU;ew-Sy7MaI4N-qp"ܑ",ƀȗnt7%Z_jpvufI-]ژFȗ~ |sw/|?~'is%Xptq$$̴ UȈp "_(N'\g30ldCI@<9'Aϙ82 s/?g\+;#{ﳡ~ٌqUX bN <3GKQ&~ 18[Nׇ|x U񪨞a?Ͷvא!U7%z>dO2VA.Bi=k <o]30ύmgRhuCey[Ήo 2Z2zNm9zJ!+QARp̸t$$刃oSv"qytC/EWS $Q=](찷;^EG,!Hn ߌ6]<|ק{P ><6"$(>pojE0Kq]ݨjF\]  ["$1@i2*cpb[2!cp<"|!цHH\PL Ƅіx'L9"gƑRH[V6$lvP ҥ - œguIeG>HItpb ̣W,2![l(躆c50g?q=ٟ]kt95%@j"Qegn-qRzu|r|BFlj𺺲̥!?#A ;_8,]<>gD30 dxBn@EsPT ˑ.6O_*X[ \;_+3ZVHL(>|)K{L31 Enc0#.ƴ70 qگ96d=qEع#>}Hy -!xt^E/EtU'ꂽ/V(޵M S\N.Ab*D'ϭ+)Dsyd PqE!ɻ!= fo_ 0c>}/I0E"#\ NK/>?o/&~e_oy:~.d~g޲99/x=xڨ6wS{ _{1](U1!gat;|\uDF\s)<5B{]l7aF6{mDKY'*^ry@}6`kM5C(`E/a/GhgWʾ) rO}\+ӎsZofskGA)oi8KG2!c(BLȗBްUͭa7;v<%7,P6Y3P(#.~Nub.D/DW8m=(~@팙JY 7vFkIwBFǍDh⇇-˳GQ^u;R8nw8e0 bҵs7;vni5nS !WpKpcG-YEU9BȚY1R~vg6&Yυr,?4@A=!g(ռmvg^k2s9A~J`b$ LTs; ߤ휤c$Z=R Zp"$9EO[6c2pIyr1rE<_ ߯hfٽ[3.gA6, NƸ7\bDZ`eA2/H{0X<=K^B-Q-]g2!c0<&:|) n7ywAchS;YmϪB6^}$ߵth?%Nf-D`>fMZh)&Xόw#"_YyqVѵ~vA EjKx')" .2:}ot|!uTBIDF慕`bece ⡱@/UnEP"V*pkS;>bbM,.V"7plZ+,6퐹E<9ܢȗbJܳOB}TdzRsUa7 &<@Mzm0c&jׯ+-]7phn)A["2āELlnv5ޅ ``8*** Sk0M 7;V1Ku$O%z<¼esCʬ 'ߎwk.VI3Z{ :Z/;Z-qEZzRso47M\IfTxLrqBFǙF-hWSP{@'f, GKQKmMITn\vz8C8M6{bdX >]^'C cwB/NW0|^f@j &DZTBT֖3予#X8|!!e4( j: +6lAD˵!eޤ1Kqf=I{M p􎾩ճWvBn65 faI4NV_ 'k=3K ~1.cEr[_kb2{ꈶ)$!B%RV~1CC/\U$/?oKT{vnl|¡ԟqc︥_˄]q95Q⌱ e3jP5 n T2N2+clȗBnMv`nOۆ|Wسjvo팕2:OLv8B;+cg4hLx[v ' }7/Ÿ`tH/j)J\cO[qC_fQ#'dtcpclW j;piz2+<& 6tLw/.x4v|X_32K&×\RJkJ& ^ +A/ȞTj{T:Q_q) .a $Y>2EޥǼ3a\h/=M6^qQ [+}1EӒ8!DKU;x/3_!Sp]<>$< r9G#GȗB`PfH BxOݝ|ʵ3)5t,JQK7pkI`p gia\{㸍俊1;rG %C^6A2-(ZvduL/ݗlK%=MsF]bcUd ܞrG2`:?ڲj$hMm7zl'4_C_W/W7F_a2dNyAdyEYW?u]ks1n :y\J񣒦F3tQs 3P Le­ծB(M!HGj;eQ] z.|%9DcqFV-mdʩ%"%9C& NQޭyk/q~L2~=88tE\H3de3a@+/cbJ482wE*쵖)?l3V^V H(3g!iQ@Z[6G+sk(c6׮F}?6`P4~tFV?7˜k1[>*xNӐk5L5ޫ=w2 YsbrӌKeGs]}(p닐*\; Fx|р;LWm %>Ky"H >WF{| kU)WSJy1sU՛|ǏTsa˛K&÷]6 [ٶܵo+WW;/uU΃DkOKe~sΠz^jr9|7c$Mc?vy~#Xy*~_r~~5C1SDp']dz{/~m1[܏ MûnM_=jQtٟxߟ/ /?7 _{)LfW:d,{:|_MxG$ae -Idl>/+2_}Q Ͷ=/Yf"\>WjOG Fv6}wr3g/yy}l쿷s\{v3L.fݪURo+6?>=ng?j+K_7 ZvmLk*mʕ7yrƃp]pکqiվ7\˦Yo,Z~VS|_z 9O\ҮbkLSů}N_z7 \"݅Tξ=?sW]{BN^ ^^/sb3EHLkdse(%ݟFڥZCi x ]`۾gFKM#i Q6([C a -x SEQAU9!(ʳ47 ij9KF0l㸂RQED<.qj8N鄡~ɳ̤2cJ 7&-roW:do| $H#rCt1TKreBwyH*9 Z)&+v3cG:ti2iҒBy-N 1|'bR$8VNIlǘAahO5g%"y AZ܂"HsMb~MjgH3 *a s?5˽iay9ʸP$6 PQ|@<gG M" HE5QwAZ߂"o{|AXN=(fWvEE}`^v%x$:P T#l-MҀ]Q%\'9etSP)I0tz|s{.G/_x(پet-Ƣ,1mŘ41im s&MSZD[ie3Hm &xtj"SMĸp2CmhMRYtse+A9j)Rфۢ "f+CgiKgi;-58ɡ.ԛ-onB EcDy^wM)򦆵\d׎Qnx|<ez7 &h%X[ ~b bL2 * yⰧ2Pov1xlĘQkCR{B'U%c5#yfj яŘB#AY' dۧϙ'#oN6Oz"j(d SxeC&QDP ,7ȏ̯L#GqRJKsiq&zn3HDeGu*|^8TQ{Q%GB8n41$G3I<˜9Qc')ᬶ8GwƜ1{SoO1rP"!41LK3?}Vpv+($(67aؠ6-l b6杖p3@`p+x:9Ѱ - R?9!>~&`f.U}n.zSlW c%ڏs"%ᐯX ;*uhmp.*.k <^t+xY_ky }v|+Nə9 -v{eaAJiLܒ]KAA!Lїa1@TR˖Gס|4u㡞}g} E PC;բ@Bn('^<W,Zڄ֍F xxi&'Ǻ/ &|/pZW=1[ekS[色ŘLe1c(Pscy:z DTlsʪ`D[ˆŘQJ<$ԛEТw<$ :(EfzrUh@@G'׈e\U) yrxڻ:zFa1r8ZPԝn],q]ݰ6ȑBG~BDD@34!O-ua1AT* 2gGہo<Ǟn&4˄ĕMy\ssY[{]IMn;>ٶ a1&lU'Bfxx!;bğZ4b55!$^1" UR3s(H%X]nm(H IC=LPBq~Z(B~w!^fXƚ$)ΰeixn4!I162K8>'@Hk!>V``ҧ=m#-utc {r+ t:0xg{҈y@ozҼOQLD[D9U~Gq;aR," {EF*s)RJRhA&`-O+SH1dc[KyL0 عe <:B|eD H,F+ŭAGw$A/Tx|:>|F2. ^~%mneZBgG/+:ʈ1~IxX)f`@an|gi׌cDQkjhiTF+` XݢW/av'[5Cl>I:؀kP /MS^_?LI*Q$j@G˿)Ƨ 'a*4\”X/,_k2{b6 H1gCM̶4бXM =S*z`?)f%yP|q5&<3Ëfm6p[lOf1]"SƈG 5dgbG춫y?rv̟}O޻aȦSSzUo{ԅ<ʨܫfa>n sMG$WӤ'UC]o8%_tt$fI9$%$؃#9ÆtҒp[2gxoh/ZEԻ:5Tw,L?ˋ8eb RpCH`fԢX%WmnDwqnfME4jZA]d%qEo`6qY (N"IX^̮f(j`5PrSnEC ]?G( |è 1g[]G[Nmό(Cn_mLΣxR:QBpl.֍V>Gq4 K-jq>1x*> IN-O:FS|X&v.8#rVL *=wCgMU?1IK'86*9cP[Q=s*9^@ 6P pjTOCQ= QJj",C"{v6n~ؒkF!Z}?Wn^+]6oyK}\\h8u5"ǰ;5Q +HYea~Z}Tų6^6=*e@8l3"L f5z\TCq6*-u3j15ldD1\X 2՛?QbCV_7?&"迌WKX?,Ѷ tgxCixNd33fEk_P۱ӯ7U-@^k_SSoنAnYIŷ;;< \ E6vBeLZ/a{8,(h3;Wfx-*ȶ*y6ЫڼԽې1U:=*73_ s` ]"_:0/Rå. ]]O%6\ͦG=Wlh)AGӺT)-m40)`.Rt9pG VM>ŐǶ gT{#FmFGovZ| <N,wݿ%4+c!#öD;!:MIf20>~p;,9a; 1ӧaaÃ= }Epg b\Z(Tx؊EcİLu^Ie12 6 % l56v1ATAjgSu6JSH^hJ c2N+α)\"D|b2D)3 I wۼA{|5Ы ŕSSA=ɇW'܄~w2І/5H 7{Ξ!.%2;=v[rTтt=}?Ũj4.Gu)XwrՎ?/WѢ]^E$}T/VoF57uŪ'8y${;{R_긜.GxF@$|4[|yt(s:*-za_\TsgZ$}'{U\Kd'pJu?t y%,/DT 9(%uwd*.;J栀VwHL3jtVSǑfҔ8'lTb yzzN|'߯VڤTZ1}p.5vR]adKU-P(h(.qEJIal\n6P~$a'JWK?O;/H3K:cwч;LK00TQ.7NHXJr); Ǜ#0sbֈ`AkIHÉnދ.YsiWZ_T pҐ1bh>WZl!xɑ*! =ܖgz7Ы2 eSd !QEuFSR4ߖ|M@gM0kIQn`i ~\l2.5`e1 ;Ul͹L(f/g 0[ +WOXɦaex(SSMÆTC`哨rk4*᱄F1’#FVIbjΞ0^ۜ=8&cRQ'Jh1#7ژ&=ooe&Cc}bٰ2Bel0ArޫoҳN*LVAQ7 -N}Z *:znٲMZf08\m_Wc0qhm+1Csqbd|vQgI[o5jnS3i@auUP|@XBr 8Gk|׻=Bt^{+)7>)C%  )QBG7d0[voUb&E/%f`20la4 NWcR7 ^xu͙IT;MS_٩}Aw6Bf9Tpu M:^{PX,ޜ*.W]"x`??)U"ſ ӿ-C$\i‡ET>+jP՗yXvkeieǠs,Qk@jlG8tۗ\OnyWs]Vovy;jѴuC7~9}_d_{`3[Xn%pv9ܻޭ_,gsY 4,KٵY.UK<ۃ`&@k@>|:s?奥BAr%BIsn ́$kn0;X59/"U4o? ǰ,}E[%l?Wd\^z_֧ ,0Uo_$gimzTl~~=LGsMmg]qQ%+eA#b!7wgƎ[mF|-VD+kc[5:սnb#kt7jqb^,ǟ/ݯ㟿ur~T&-Ty]2*>{\뇤w?]GR8[ow`;vҔ& +mIn>j^v NU"ףADU5_>?'T(UpϮ7/#"&u hX2t{>.:?:'P|L i ӌ>{C?#9`LcE%[rn)(zrHB ;qJx:G@r#ᓺ]k?el l=n|IO@;@?$au_u k;վ7gXǼ1~9\yWL\9e&P y@JP挙O8U'7XmS?r7-釯"}7%v3HpbJ0PuG tx@`NH) Q 62g{szBs%mxQL틔}q ٓS yD_Ζ{" 7h jg5o_ Yg~ F :d`%<`EAXSA䙺i+qC$ Q{sC&L pDಱYSro­+&іWLkZo !Dt2{2P\n59fF2x *.CfK|Jɝܤm>W [gW߼[ٻGa-L ΉIwދ9e-7(N?E Z5!kMĒ8/(F8|i'G 槈 ydfB7m7u]#/Ӣti`ƾ/vj<}NAXeAc*8 %1AJUƄsYKa1h߮#/⋱xXXg _.srku*}1h< ݷ|uO7IOOZʾSZb>CsƟ=<ː XI<%Ƕe(@<0 (PR%KcUqd&`eꎉJ̄ydYH_}bɯ݋7fv/u"ĉ݄`a`Jwq2(}$81CZĽ?PQ6!Gb / V1^S0¼g.`)Dd#HX/w }0${3S Nq? !FGOR7ց΢N&+7q0e1( ؆K!g J洐^֥`y#U̧:| Ӻd>C/շ6JJREPbE,2p?#QҲ#W] >'RޙOix`{;m)FS03eTczu"L%FLxK栫 5XxA 0/y>E~ }0υ/6^/6^mtxmK\̂kܻr1u_qoػFn%W`_39@8 'H/{l[qZ[lɲljl&3\5Y*օ)]3 ^uu ₡ U *yUxk)SwSeQR0Jj B_T뉡Bx{_uP ]ޥ eqZ3aE%?ePH"|<ʪ$wQÂbXi+Պx )/Ln'ZQ,-C+xoR£W@ )Vx$/&e((q @C!RzY7U8mb Bp.^H&@m!RC (l@ X4•0ih/MQhq: ?)JG-V$Ƶk~Mx siN2A[1Ml c]YЇ?L!H}nJ0 XIdp"cQ~PH!&u@ ƥϊ9C kPX\`($^|1i|CѨf"|ɩ)$Z{iBM=@2\qWU,Ua}{c l\Y˵Njon$ʷٓwEqԝSiE}_ޜ!_ R&ax2o{wU1U'/bW "mMXY*$W:"(C4[j h&4EH#‰RQQ0R_0:cKG=ԁ9˙?~PH"߼JʐE$1fXP"˜JZ'Bx+v#A# jF9+~Džvle3ls?3x 8]_ Xx ram~Ih6cEw|swب7SqAj=PTߎR0jg,`9M-(×G^p8iH)pC҈`U֕ S?jIxdQܭ Q(~?̸Av#ҕV'(̎â˲*L Ƭ~ Jp5MEKWMC6l5XW}i@H-{0#$3B:AI*u3ڙ0d>Ψ4E$px&y=S4zf@ҜA*r2zH]}G j <I]nz[< 7쁃 ;*3Z*AL0&\7򯫻zDv2#38FUaYrOvA2(Ӷk1!in̺aq#*fnf;E<0eYJz PJrZl s1/Cxe13v rvڍ N"/gDϗ^^>"؈CxdDk,XCoWC$ET$WDžu (dSdGfpIo7c*!<򂣨Jo/gg{Q`2GH-ѴN0I: UzlLE2MLQI>}H1s*;ir%:zH>;. (v֭i8eޫ8*G= ^P<)CMxdǤʚ#ꈙ׌<2A]J+X44X^bElĩgluCT:-̪ /DF[gp-\xLPd:<S|b,<):Af9_,a/Ca.7cD]FPca",T2XGO-:lfrɌQAm>K3™Ms8LFDa)[G<`*xdǤ/z9¿L`"<2S(ob?j2"HPB2kÆF.UT`D)x&@?#38 ڡXB^vT(9X,IHT{WnF\y$-YhMFJh,/CYa4gVc(m7b)aIPsjGwF e LSЯYä ^1N f/RiwDY SX(ue(v֓/2jxz=3I H?:y+3^EnlS5 f<7>_#BUP%H%) S3_3 UDvf0NUe"iQ%]Ñr>%i~f\#p>G;gތ^ގ(C N~(Y =""y ʫsu?5~jG^pFL6I{^WnGDA<2Sȓ\h]=A2TntMbG%vG\az2 K2mymja.7v8>I3RZ3"NO™7 sƠ1 3,J$^gfi}~fPIZ:N8M &'zD8E.B0]_AirQxd'7K &zGfpdVtWT?A2PpmJ0^](Aո hLI64 ot~ÁNy Z&T{jR9A;wUfƑSTT5-ԕ$WZ6bs%')M`ϻ&,2 s?̸$(@=Wq  +sfoYOqꖕV'$S.˪0N)O]6]*ZQR<2i[\כoa(h§N=x'B%MwY{}{5obbՂSe84&)k)] (CX"TjD!*b+BQzyW]_4clXD,{Gfp8tD}I_}:W#:Mvx$7@xu2"a~Q^ipNU Oł&ؚ Z)^t|N1.q 'zD*mz)FQؓ2Auo )zu5m55Ff Cg ))D Ň +pP>Xf[LwY;[–]5E8"[5 g<6g={pup0DwggO@5O_m/mݓgcZl/]f grYW횪˧:3ka I'77O>;]Whn|ãe}ޘ~"f-rX=H\Mi?ϖ.{Weyx/VZ~xtqH?w Dp^KVϝ!S vj4@)\~ha붮9+<"grMmFCDu{ w>y8ghEڏ6F@ُZE]/?~5RD?:8{FϞ~c-zۿ f7qM ?ة )TJ,<&7u1R_g"|Bxdj1^C!cR4+/d>cnS߂ursŏ<Gӣ=sD!AsTP~aa~÷q =+r!dImI y{ \%箃p߫͘V!3:^V0iWIpq3=5>p| &"qC:&h`0KmlÔt̥ґfeYr:L/TjN$sW4-g7q~Sv2>E~,=!R4I}UdR,?]T _9ۀq-D@뽖t)YK;':^k <-ƿ7_Vㄩ-D|LhH5-+Y21*/X}G8xS>n1lDf782DpF^ BN ăP%F0&]=L >ig8A ĩvqkze& B5c ~&Pc3Z W  vp 0)ܖ (8U(n?I}@?Hbł7d hSh_ހXW.qJɷ6hl߲TE%wv{s/yE>ϽX>ϝzR(*KO]UU.;RHޓh9UOxِiѸBar=#)lRYE5,녃[Ӛ|`+^@<䨨~ָ뫣)|67=}Ԕ=]4{LxZ]O#S}:?SRHT)Yi֥ZDRvk,j'5DmzW_ͼpuS)GV^T̏zBJvBpe6 eEt֐J ԉC;xSpQ0 H# 'xe3T˨BdLE)XDfq#rc{`h|;r{7p%e5Dߣogmn"箴>ug[e58+gg"h9JF!ֳ(NIf(i%hbm:_uq ,M\a:@Qɓ0]i 'Bzk44":P@2kPx>ν>[tٰ{yaʺ?|E)eԄ rv$ƨc%QIH" J}]r8H*R!ض5<&/+'mLM'mHlyc.(u_E )}^T &>-Ʊqu!̂K|bE77^Xvj2oVWS ϛg띚Hij?F[UpPAϊJzsIesdϛƦIΉ iUa}TQK{^k?Qچ/~cuf{ivtކλP7obֻgtS6t-=T{*ͲYKGQ[6O"dp"ׅ>15*H(+rZ+)nW~w69vEFZؒQ|Xba#a,zcp+nZצWoj,(| (g|/^e󋟿/_]Pf.ξ8xs?uḵ .pP<7ݛnhڛ5͍إNG=]ڵ]/Zٍy? n |;ky^UTJpLHS1sIT%_9&&W:Ѳrw{zeg>G*B`yj^Ֆ;g{fDkEg9&T+IrQ~R0%4 0xd@,NL&VtsſS| kD{L#Z^JZ$,]_"ԓSRuA ހؘ6(" " " 6" $PL3*)aL$jH$jH$jH$jAj4w C$:HށHށYI(ADlHށא;1#IQ; yw B$@OQSzP \JJyR q u!x1O#P2ΥܺH^6 HJψXUiʹ҄'j? GkZu؏GaDoY=upL:ma nimimim! ڠ# 9ZFo ! ڠj# ڠ@Z Ŕim0&CZAZAZAZĴ͑pa$9)r8g5?`'OZ֌b5ʒ/ĊP'Z:&Jb34~)<- WG o{H}W2 rj~y!֑; s3K.2|T/*@g9V|6E;wUbXQ>䣓)oR7zȐYlj#5kmgͭF\aSu.n_5]~#wƈ|7U|4ݛ-٧Wq&G-jߩ9:]ZQV`cXi&DWr|e&M:z2\X x؂ `d ` ` TJb2"bC[dE ` (µQ`0lA2"-2~@*2"B[dE[dE[dE{aj;aqC>jS^D( +Q*ދDLDŽl[܇P'ʝKpL  07A%CyQ!uYunh->(ORL<*b, K+ͩu]^5ЇTY2_ziEM,>:GګjWR);$3kk\YAR'2%u69ΌQ'%7Jj#$Ɗ{4&=> F=;!B%UIMJd- K)wt#֚q,XF $2+3 &Q8;ֳκ=氲[~(@șSRs^]:p۪AA3D$^TQTj֨]"SCbe8%t\%MArUŅ^T'JvRxB9 VgPO'aƁ`odACb`5*jV0oG|Nj\Zo;O*x_k4x s,`/O4uƧd(Peδ\zP3Wd'gp;ldHޜM2ŹF*_)əDY2&c8s&"L;T9+Cю$d_Pk7Aٱz?)G#M93d sARQyݷ  Srlic$yn6m>CVvOHyv^yZl-X[qȶO\f7'nzBGo/sG'G?YٹM7$kmXe/7z~T^;$u|1~ʄ)!%>Dɤ,K#ˀmi8~LUS d@u5KssGh\==P+nZzVwtýs ݼ\dX7i7\V +9eP S}}6?ǓI==퓦/_b;0 Oٳ߲onuM^E AР_F¯ۢ%_%_%_syDKKgDKDKDKDK'Q/Q/Qi/Q/Q/Q/=%[#Ԝ*SyMUT*$TsRoBlT*^T*b*̡T@UT+BHQoB@T*SoM7UwR,IVĦcCIJZoL 66Ȩ[ô  '6wl:\1 j.K}fw΃UCP, vHA;YL)E6τGJ:DPp3<  N`]୷p) :$xI%;ՠ|:࢜RެjIG[^˅'S\|?+G iS. x[Г}!"dӟ9$6O+%4BKwE@ga%t,:65ԈQF5c'P(Gy]}$e]Skq5"+]Th,ake4~,lrP.g""?;EZEXD5f.q0M2I:庄d-:ګ %IU$rtiD@͊#N,Tv8EZn8C8MEQ͉BkrazOye4ji0HX~򰴳ѨƢPePUcN&'dŢ忚ϻH%Op8BuíqkiGyϰgK9C%VϩlO? Ʀ))\$'.eppv>ҝӓT@a=ƶlm 斎5#7#fօ5*bʣqgZOGóDN>8nu9mn̞9^HmBX`.G|A|?zC9;c]p +G`xlܬ伴N+煡3\غ{CծzF$/_}W}}uׇ\ɼ|.[w_PGM7-Y_Mc{˦4~v]CF|3 G֢2BEXɼM1e0+>ZWdRQ;}L֨=E;DT$_zk΀k}t}]sg `9s5Ǭ@hA2t_"$n1%&<[ߥAs Ix&=awwIl?CgTsQ 3eڶkwm );(Kqd -5E{Z=%HB7>&x`.#E-[-q \ANQHYBZ # k*?|k2ߏ:Zk3V}e<%~,XCx_-FʶTY82x_`i؀f3W : 붌h0j %rQzzT8̡k -BA-dczU !xi>k0+e:2)N d1tYA~>G@V|t;wPq0w"-ݬ>ŴCkĶ 1#$FH!1BbHZ!QfeQfeQfeQf/ QfQBQfe?QfeQfeQfeQfeQfeQfeQfeQfeQfeQfeQfeQfeQfeQfeQfL LKYDYDEYDEYDEMYDE(s)'{wPRխS*xU ¢MJ#uP*]KT7IdHޗ^@`XJ&R٬1x6ek 1i4Rf 1'G;Ww>6tmeP=6[l}>A-г.MkJ\R;`.&e O^[|/u{V2 's47mLq4leouJEd U\v&.\U.\zRh86Aj/Ǥ,fXC;ٳ֝koSBP]9Tu\x z'E'ɒ6h[ 5΂FV@ƒ7k[=^wCEvEz*Y40 d(t޼+*%-Qg]LmX>8SSGX6'gIMCk Φsڇ}AxK.ny G; Q,$Y-gAVΎ//_h ]6RG'#xJ|=0Lh}4(]F,\*ֱ_V#y3Kk9 B̸爯?գ$I3dɩMn?9Vyx.=MM U(eqGkr6nqP=4u\Jn6C5 `F8(+6Z1=N} 6B"hA: r8\=CŊl_ᵯ݊r' y\!eڹr)%ij}ăjsOɉ>2d`\a2nO髾(  Ho=0U"D AbG&[BR.`rR`aVKP2CD#Xhe :@@b>FPT )XQ8W]Ug]|4D_ C>蜙e'a}d%EfN7q8+)gk30 ZdbD{>t[w>}~Ya2َ*cdy(W lz#ʂ5Rn(4Y ľ" [ ih ?x&cP)K)=cS(łLR٭Kl@᠆ꘙWƆ-~M:%rbVXS|/*qS=^Xv׮c렬s6)0 1Qj!+aYDYAJڥLʤc;c= E4O~73?Jaa%я>GW=s%t'͋ Q޺w/l6.e~ O֪P"(w3^w Lأf敛v5> =Oٻd-WQ:Vmo_旺ճn,>g7.`أZ~̧sr:z 4/~3.5xeG:{iygUWkAXhVۏ/=Q% J 6"D0\mz-ȾPymp% wx;gl<\a%5[ KhVh#i"_v`d y`׆ѧL:"Ҏ_VQ)H1`"Tuë߹1<+Gj3iy?c6UO-CQj"ZP]dC3[L@׳sX#FX0dv7Oz+I1Bw^ɇmr 9-es07Dx1Է72Dc'[ w|o.(Yt/ҾȪLۦVj/E1RjW"FnkҎ%JB /)Ef{)P7օOOdYxVJ*1e0d K;rn ]mBZHlz21`[]Pիb- y9F- sNI+Ê>+Ê>+>LǓջ6*_VOշl^OxJ:[ @F*y;Xv5Wq}r-q}r)-q=q}Le-q}r-q/hݹ;mԹӉonPPGL(MX1Q\ѧg{G_zG0l k3rFT.!VK}W_aE)( -m?`NSi lh%H ]?y_= EΤ,* ,L6_cHL[DV2$$zf"岭&ƚuƄ tgMg뾹Wݻoo'\ґa"WhLP}":HƲlLED(5)ɂO,% |z.?[@ J҃)$6mM$cTNzQHPP3NtҏNFw b 1D%ؘ*hZRXh42be=&֣ DpLَbcc$K9b]M'S4 YB6jZ&g%@6}ɡvETNdŒHѩ^G!P{Â+}{ 1] yDL]@BJV^ MQY5"5d569`zC T|«ZvADǝ+N%Af,F/Ui RI-!vH|yDd/-@hГiȧ(AK 8y.[Oݺzգnɮ]W@Wm<>bkUH두ѼʅjAR I}&}gLl[t8}[U^>oT<O*oq՜>oO_Y"%^ 1I\ࣅϭ_=98PݱՇ+^<:)ч]ɏ~bm|- a$#pH׏tiye6J,oS>9me`*OO_=st<޴8ؼ8%nuMnZ+(.p d/x\!ޚ :u7eibŶ8m ~Ih2nnO,9GjU"fܝ_aϋ?nuf=[\YIa.qan 3ɟX?~|/{^*_+e2SNPz \~h] - ͸7qZ :h_g-!$r(ʁ. 킲6.V>:*t4$Y=k^0]Xk^Hm)#9&8)Ha5*3`3 S (Hn"]W4L%y4XgeH>N+]%C}to \4Bq¸4oL:FdH;eIeWqeB(CV'\qD цrd APNalL m."T)+W msG&u6a pGz$Q˱5ݲ>駓V7/` q1묵A3 tNq`PHK%I9$Q虷杣8u@m(z-^'_t<,WZX&hGF 5PF0 IXዖ*VTݥm=Yrw_ xS z\Z ݁$AW~h.߉onPď:PHc5>`+de=MPQd)&X]|PVF*/QnE9&kBv H ƂF v^t5`)NeTT{2GUWR*C/ZE!E@Iz*#ZA-Ğs;fcqpgC3)$/ ECĬ&f) mDbڒ(&ϼ` !!y,֓7+)m=01֒(\q9[gL ūַn7_}x[cBpȷuD~QNBB[ƘchIKVԖR2| ΆN)Ifx)ݦO)CG%XCTiON37#sr nt NJHu$HE#@*0*?{MrV0#*)dFRtKdi ;3AY ʪK_AеoEg #!-<WfW)%2RLy_ZD)PVcl;œF(@pIWqƾ:ȑ#0FfhtI-!'nj(f1`|}s{F7+dZk/X +"aHZRPk0ΗpF:< E-6`MC7hg_`@jEx(tGg|)jʐLF FŬċM)Q Vk3lZv0V Y[JUbOŸxZf_庿P=xu;8rm a\|nwc}qtx]U̖Tً(١Nw`զoB2gexPQC{a=wRf^=1~A3Z 6EQ^knH7F8 q^Ini^Zd1?u*"e9*Xh8#eA3>H͡f*&iQ咉gBEVcɷ:NTپ4)k?-? fȿ 3-dJK)FDOl0)Ic-sRL @(WNA&T.4!ʋ '"팍6+L29 yAKI0N9b"XZYh6dIٚ&MN+KggH"q>֤&}VyfAczyl!+|{\Of,9"h~Yb#:zlBc jB^`T7_s^ݹsª٘U6PeooFUV;/vhEѳ nTl!gSmd.w-PL BG{,^'٨c9VSk4haOT YpQը0?!SeqOQDriFh J(|c`OÂv+~H9Zn (.; VFhd%u|p6ŶNc8Ҥ<ޅLKxѬrl^b:v+9-@dW5uZ$H^,0T;:vSDT8>E~͇"L)nX*D.9b||¾^DFRE"郷%U9CU_nc%>< Mj<SX?܍Yk]zrCnRp6axxw%} ̒s8jvŖVC7F3ՙ{~3Le+;::Ixp7!c1h}DZf>}3Hki}-0-:50kl!Y(/|}>d+=I"{I=ij8w?$fPW`̸ Q%@i3Ck{ cDWmQM:UsL!5dw0=Ӑ:Ȥ!1: QjN!Uj#^C/EJZ)o!D0= 9+kN4{}Q\.LT3 N/2wŠ7N%RVD\0Қ X{} 2؂8ih>3ݔUEFOzݐ9x5m[מJup 0jb]Pbk߇&)MS\$A2(r ~T%ĊqߪUi6c1ɵ>-ƋfɆqTP\volLZepvܖ?o?aᬩզRᚬW~g>Sqt1(h7{݈[ ,J"<cSqI&}@t{\ɴ6αpD}4Ӊc嫸aEnalS"5:FgLp e4˼D2լd In'ZA۠F`'$A)-gv!)fיDxnD%Pi c Lr{={5ڶr+' <ň ͐vˈȂFGCl+R`k477i 7zvQ{Gb#zP5AK+ [BbjB(Jꙁb"JF* X1iBjO4QAJp.3 aEFz95&~v*<>nֈ!!'Fpi"5^ch`[%YwAYR"  "ѨIT{b80,gT+]T#"vqv3s@׵"(RVEIpiB$4!aU?U&@tAɕ.>Үz3dPix`,ڤx V0 `ICܐp(0NW#kX˼"DK)UK-" 86efJ#6PuO) cW"Fؠ##)BCG;r ji4 Y\t93t)0 5L $s_AN|.-R : RaC/6jpsfYfzWdWU b׀ O t-C6H(HKsvXu7kk??~}"Q1G EmGБaYqG4kҁ)$1ٚ}?o ~m?ocx9oG9m޳A8Ȼi olqEST|ׁi.'/(0kZ]*gI>%R5*mV@RR@AY,|M4zC!..ր#ĨA+$ `i(|W.tB0Ƃ RYD?e*0Y͔WDhL$xX7aI]'Ƽ+L;),% I(..}ll,OƇuMLj9HFXp `Ns=goT\z[ %˿{P1lDԣTQ_sO -*H3}nDBYwy ?}.歱Uˏw.9rDƨTR,.8i2"QD 8_=zޥ:xhʹ,\BNq?y4ZHO >N;qpQJ0[S!;T,ԿTxp")9T+!̰4iG `~SM/r jƜMk4YzhcwmXq!$OP-?dK}=4#XZ~lI]0.|nqxV-Ջ%@Նo.E ;g 5-aweKn|]3d}3Dm,} #F1EpɢyzStsU'׵ NҞ@l$5A}</'b2GQi-XnqR!{sпOIyO/B?CW2,ߟdn޵ *@Sߗ]˯3׿8#ޝ6~Ndy4IMM֏W/1:g#l$"ֈ+ݐGu 1NJs'46y%5[uvX*rqZ6V&* E< u1飊X9mB uX҉GRoq~e ؝VL3Mv~iRV~öp*Dh6WH8-%& d]-ı9l)&am ~+'C:A&T.4!ʋ '"팍Ue*zg2DOL^aBp6Dd%0l\tc&OeW#T6q a=a9\5e؄|:ry]Of`,9"hF,|ux\E:mxg푔H@V#[3+%ġv67F/HkB9ٲa6NZa O_hZ>A`wxqND"d'L޾Prg&`1vM׹~;3L͗^6_8hڼ@ݼc/vLb(x 6C-,vKM>/.Bs&jE9]nTj\GłY*.\]Zf{;շY9$;pR:ch ;y:+]L"a^ݶp~?e.sQC^|WYۛTE$s9(\s S]n!` }:h"WPhz_r06)kCXsH&aqVƜp~1-{)~Q./ D mYQQeMis4JIB/?4eH.TssQLtxӴ8$V~\d[q6 n'G/r>Yjurc' 4`RlҿuMV Q/6^:Ո"!PO,n?xC)e0c>*&ɿi)hf Oߓ4z67yE= LkȝJH 1W Q&(Z?wX;l+\PW8tۃKn\2gK%K`\K~DW䥥B凛4)c:x*;H:EХ+/deP{ƻXM&cI'UAC>]w8⚗6VJL>dĔպEj0=K} 2*q4* {~q90U?>cpSAg66>g.~̙b'ɥTH~2G@ka"Y`"_}Hw#5Jݍ䵌da{VZ/g[r&.{G>}Ђk: #; #X;#;; #; #; #; #uA~GwA~GX$$BXvA~GwA~GwA~G? 8aDنb|N#nBW ҭNUmM^";\iKټkk.PHx3jg>M(RFXp4@gؖUy%YpĤ6d+~9 %ws _ &PX{Y0~Fax5n˹y?4}&Wոm6xu*\A[:_pҍOsw_-ӿ-v%'>WZT o6lul׈zbA<ûF{LnY1^z@/`k U cJ#68R!jRsQ bkQGM|LJjv)CZFD iUӤa `\8g!?kby sR 5(V8d .Q~G0Y4Q)mD0XhS0M'F찵ˌFrДjZmL䮄Rn&xy:, "A J)#RHDcۻy:m-oOҍ'6 կK=Jcvf;U, A{>T6oeV. $W{qUXfŁCc崥}C" VI& %:cEH, +nTDZM87 #$)W &52(B;NxMiMi hwlv`QlAEu @l#F]1qw =&ñP~fu Q yp)m 1*fDEXLڙ"dBJ*\@KeVfJE"JF7{,ռrk:TnƼ+ VQ>5D!cIgAq%JK:; Kľױv߃z@[sɠn S`f>ٚo-'(3 Sms,b!!vV'D_}lo,aV:81Q8K)> Ptg f\e%%c88/LPҠmPV# $0 `r sE[lɇ ;/8^jl@`2m*]xKe"B32bX)J"  %Jsc*]msI+|ٻmS/D,713w n ^Y2խWےeԒlDUdU֓zd$#y@rR==%|<FAdWM5BF"W)wuڙ#`֚q,XF7Me&B(5Y1٤ĉc;;[#*6nPT|ov 9%E䅛EKEGN[5έ##5@< Md mAԓy dE}Yb$0و:dpA`oRm5w0gF~[2_`/yT8RaE NȰ:XUچ ^NrI/,(g+V_dH{#̾"5a q \~mJyխ2Sܻj%(m 3oDPyP5Rs#sicHS3{2=7[=2oz _\OӠ5+/Bq /bI'oy5}p9{?kۇ:I,WDF8M^P*Ǖ_*ǭr*ǭr*ǭr\eUSqy]{%ͤ{@VA,s͈.VB#z8js ,DYj ]˓-"#!-M4eF=wXC\*#jDA&*/0GwC*Mkq4|9gX| ps߲?a'AD]"%";`KZU[)x6{>QLgWNOFI{kEߡ50g!R-]|{qkA<֜RTX kcy5\ÚXsk)by\9v5<֜ǚXskcy9q֜ǚXskcy695<֜ǚXskc6&j.<2`y95<֜ǚ󈎦HB)@*HX  bT,R@*ʝ@RrAt1^+Ha5v_o3Aa`?=BhTr ]VAډ]WRoIz^kbfZ_;:dg|NsVu-LA՚(-51h:^;#uR8RC6lNIHvBpU$YC)hԙc;ƌLI1% %CײDw\I"d9JGF!6$RReLV1Åqm6w;¡^%XFUbZKІfPD6"pkt]I490 l BNQ}/:wW{le#ƺ%m4%YvF%8ˤ 3 .9I)e6*yMID??^1QKJBrI6c0Kh& ȭs5h  $GC֐+wɤ:Sn9+.DsjNV׏R DseI5,F)QORޅb s?~V`տ?O/̮4}ͺSӽ_ur~S8+dYpNxqB e50J('䜜j9č'NRk" x}9XO] $yA 7wYfݍ_A ~NhQLJ)UuJ!v5d*%oϞ=dlHKx:OVwA_?ԈR}3QtsBJ5#jeLӻ 6uo\xq1028&^} rw- .- ~40/?\4x.YӒ"bmKBٺffZ̲Ci0l|HHALEd1 ׵n#筎PHiXm8 c_*[O Md sq s@$eP:;AݡWGfR!ͲF;'qjnI[xC`Jg! x}ܰZ &dJ' 'lE=JŠL܎5k=f^QC-bI%ZvdGu z}m\\.XCKsi4n6aQU6x={4m˙5sóE,Q:iovM594ϛj_Dj̮Veb4 ^KA^L ,fLP= zE'I3W^YX:U6skᬪfZ wz0|} rmԒj)` P2n(%L|OoPA}b~:S 9yOlp>NCQ \QI-9ĄBrsx="Yv:E;(~eB~q監BV8W9A̝s;K7ט}&:u)V$4y 6K1gge~NJx3`6G"3!H2#h4B~$T4kpoe7VP אD=7`a?"&{nH/DB4~T:2'u""[{- `^'.ǝϧU9]v8rH5)pZ)CTQgiFt΀2FXBs3ĺ*M{UWҪLxyB+ UҠ}eʼ9!f뤪V Bn||vѺW߾`;XhqwKy@$h礕쇳@5,Pu,*ت-o*el PHP*LPo#`.Z%Xdkᄨ=סvI7bی`v~<~ xflK49jUzetGpo2@r)+IQTKm~ *\I[aZ8>Αkͷl“G =JQ5G+ YɜyuʡūWO6 K>~WY.] ?|6@4jN&&c^޺-v: zs\N26=v%٥Ӯ$ç]#W>. (wedCmD QfXʼ^B:7Kepl;hxV k铢N]e?h; ~(w$׍ z˔ޏr=R_YOU|KZϲKKȻK-fW>w> |ݥ(q DХ-;MޥK C>/'FQ̍3Oҫz̄=1*$߯cɾ0,/sCY]ot NfFI]Dgqz}'adgo;  ~%p.C0 Z*iA;}s4PO; HQ@M'LK #q?#q?#q?#;2#7#q??#q?#q?#q?#q?#q?#q?#q?#q?uH܏H܏H܏x|+cH*\kH*H*H*H*H*ҏ /7 2 " 2 " 2 " 2 " QF2ݟnowj1Wkfl-dH|d4P}@rk?>\JJffC^Pcp[*D,ΔUJg 0$H6k?nF~5t85++׷x u3D&\06nE+~7[cOc}Amlo1"J6ǕD^e:l[TE R}&/wFo hS8ϓevYfDY28l+A[#56x#TnP? F "mF fqNJ~̴;E[dQ07v>j`otjs6+DOGDr+ pw٩xۢ q`&2+a[;߉/Jl)<Hrm>S~,gtT WiVsRL^[qth>A%{NNˤT<*i-0ֱ`B 8>u2Wp HQ!Ef&R)P j< ge/İѯc6\8U_&m|X4v|Zx_Kc'oپp&fg2v㳳YҺxn1/=Sm^= ,?ܞ?9?],8fU:! ,䫚O!a\LcʵPd jŸ"1q`e(3iHDwD4s*\e" ,|#* Λё{/( a D7ɇCARfPFXyHGrt!~AЧI2 n%U{tSRf@",d3%hXgK~ʦؤԮ4h~& B/ h5?ۓsS95c9)y攉ޕ';՛Y*sI5\a,e^%*%Gqdme`fJ#6Puo) ;+A҄AάaՆZ JuK 8) ՙ E hpjR.P2pg/>gT<xY,CjXXLjAu:0Uj/HmY&&VQR`*ƌƁZ 1b"iZ+Ip;qV鎻U*Iߩ'CϏN֒e/-ȶ*X 1B+{LںC䧤vw Sdgu2ZRɂײ&w}O+1aU&ՙ@_]+Js{5>L؛<9tepMj{<78-N|\% 0 en@rI4dtNqEtз/UnwdfZWhh;iܿ wdkl| R9JEH^=x[0>Vݕ}U|h}kE[5 cJpR6ZH)tiJr淔Ε%vw5_fN *еAzjk=Ÿi&%5Ί;|u9Y10(223iZȼ;W%u즃 A2Nq@"ݿm]%A)%Rėu wRӺ^[-FF"5Max-Z)hsV36㼭׽+Y9<uV2ou^==8_"3W:USQB^Z]&/z&(Yp1x4uG:rKöf+Ro.!kO/* fC0j )/vU. Lu!=BlMҳLL4g.ZTU~zQG<@Z=Tƨ%Mf]vjmߜ5D<3;4~<x00%:sAu> םi=u?.rN=zUJ}$%N^qjj96-xI`9y:GR:d|GtTAfu+{V3EӦB' pҴmf'T5#OjV%y! ziG[\`L[k YQLtQ[ȩiU^I-d4\:5pgg+eUavt3zF5uįD|YG\uuG dV ~7ɭgDtZ1Zw Ar4df`)LFvZurpm@~bK=MklB)ݥUI/jݨQ6T[';d2G(:ZS/+NUKPݽe$ %cYr."JY1)2{W|בc0u8iŹOM<Մߍ߂_>Oyh c<ϲo0 #|͏ww? 4{40駱Nf|?C"!^$B:ҏO㧰$d(g D]7Aۛ㘥JF:Af M{:L,3P Mfj?/I -|x}Dp'?;CcqTQ( ",;`1X+\$JKF=r_u.C%o/MLܙ)tYg_}&K;=znffo}= |o&8{N}pkDžpCD4N?OҊKx6DOn qnɤݥnWʎ/YO7 #gÅ{b${,`moa `G^If@ppڟ~Z(j](!n{Sh:}D~Dj SPM@EGgǾ ̑]wg&cO֟iK&Ξt8zb!H1}[xHܞqW} #fFttv?.u"Ky݂uyWO#K&^ X\Y8}_]tRY C2X!xmRN(3\萌99ekHkm Y/$.YLzv~<)߉ݸ yi˵ݗG.ײңwY;,2 $@WJ$2@@xGRL3ғ.#HB/:}P͸99hR'1J ɤlG'F@-Uz=$&P~lw{؏hw̺¬?-TT;'H3L`ҁG>`Pj s"fPǟ:a$XXDYYkK:% A$>I=\琬%T{A=dH2Rqt9?;O=4P0o0,Ͽo-.HnysÅ*ijjF%mXENGځjGHS9RK k?~Ql0?q2802_jӽiӓ VwD^ q5;{g\3?U>OH^aճgf^_.p)R~MS&5MiCz ]*Lsv;?}1,jw{\]FvvoI|ZYY市I|jEC4yƹi1fi#ΊK:BCs(Pʚfi⫢_s?O~l\>x OF2|?{w9[nvPD6W~>6l 6 #6#a51bQeI>?7z|;f7ٶ8yqmuMnZ̪Ũ9߇XUq3^ urv+ߋ t+ꗛO/oq,ᗟ_ 8s,!]WBăTW*ڣ!35κ/]rR5Qo%U]_|ϯ/~x(Ox\؋W}uǗt~Gex6N<4/-Y[Cx󡥅}nuճ g\7ѷƫԈKˀ&%cKclՑV=)SzESnє>dES 9W9tZ<|eF߾ v/69T ɄWhdb)2#^Eğ7G~|pxW+uGBbCV2osLY],x9H>ZdqcF;#sIX7)YHLz͜D%VYNqm iByO[-[zwf}y} ΐr!;IQ`GB+,b'@"P88>|7[5F+YD¿6{==I{v+jʙח~忕#ӧô[$4*M1t΢K %hqK۾R~Z{h! <Ԗi[:Hk͙0n.rs <%xK;m! ($L` N4DEZB1%0YI0JD(wIx Y&pKMg7L"8V_ |X$Vu/h~ a!g1S>nq@ P)A2ZN? B}6bH, 1o+;Co_Nl5|=/7aS]E\AZv͑i{۪~ӈz bSMuɶiIZkʘX{GH;Uq.(/geLHeg!3ybN_O]%nmK;WEñ*y7BdV{)0i7?̏nzA"n+ZXo~R΄ 祖vv<^6:jMxg[B9_!5XyJU|wSRvU|E__iG 7^z 0xK)g?p Y.61#QidtuZ:o (cmBKt޳&[C51Hw׈lCV)a]6| ՐȲK8̐P ::Os3N8 Zx ̂V"Y1 rUkm_{fNo< 1c"dy( lz#т5Ro(4Jv "އ g2PӏYL%Ɣ"Α^U@3"p k˰ġ:"ӯz "XbYfL k/07>8Mݧթf6y,yu&Ff:R4&JK=7&|QHTVlX9I oG'0>kzcPg=VX͙[Uͽ-_^lr^r-Z6G_>6vGP^P5`US,%e٘tm\ZѸ)xA+󰭋uQ¶.Jr֬:5HnA?8Q%Gf1 \%$"ЇN1$l={SV/LU RBE55UDj.wfjrK6/tлKP8Y|%mˎOc*djBץ":!s0*,kg: }g^ H(H4g`.YGm^ffQ{ 18'bih1i4E)GZTP@sFrQym[ۚnͭSRL+*/<,.V:@{6Lz9*7/,4T#\jŤLk+ELܻ%aH=JI+ KAƎs[ܚPͨ_w|v[HpgI&b5iЙ"^%-Npe,Z3ψB;i0$Ũ #NX4ŋc, l:dgMgO>;[k[ȅ ,)]}r^t,iCg22:#}2c6YB3A>3 I*'R A$Z`J+KACRjI޵jD]6\+JI(< lKs4X&JM|%i7CeW5JHO 9Ht: AXc$XWd{sހVAm[Dhn CI\H/z|*|)󩖙o;!C]V)i)bu21cB E6%_t8fiH[wcwO;Z/2woz3.kWI4Yܝ JU4@(DinXV64\z7-DHLt:hP:$/@B8V%#S$G-@5P"KK#G!Y29j4 @4I1[ $>2km:{d,#4iTIO9|)%7dk:d<+jx.o<;tt$tYRϿ/ҴL5QEdto$dIPgRM ZA I$XEoN/9$la!!y棡Mhi&ĥYgZQ} #*v xH9?OΪj۴ɍ޵olzw`qU澮JAee}?c|ufA-/jZ:FW֣TGT9"|SrO2h0.ZJjNΧ\bl>(h:pZBZ|ªաeDYɮ]w@W=3е|vHHH0io/MznIVD:'B%GGgBQZZ}r_ tT/h86MDq]s^sS1Ey!홳`Uۇ.JJ4^?DbwS:OR(.Y9 š%$_+>C Yek]AD ٸZQ(:ܕk+\|-8 !BJ`yL,+e:C* yΖ;”C96^hB*[:YԞ+JL""pS&dR_p)0x= ^ήPJs-'5,_hWdxT2i ܐ22`܆>d3N mbLonWK?>̆fvg52L.mJ=%L޿FqCIѕįsE56HhzDU-^i^kH"%F@^|pi6VޓT ;Ē_} g{8W!&@FÀq@[,!S",_<(#q(R"qǯ豉)ӿ܊m5s-A$ps|D^2$5AZX,_vGEwo> KKB9MroDʵǵfY/`m~&7FahZW&7;=ogɀUz==Ѥb՝x47dL۴JFH$JH1 NTF'jK*Ffd j0DVdNJlFimӫ\xol3rw5 F7d"hF` >OP fh/#Et$LEt!"'U:`!@ge_E.qVq1 A0 喀=rl4bd$cZaR,g*xHLduD*`Fbu ٴ$Nv)iZv '`ME8->ϷËN #Z, 'Ar_g[Ƣc\g\qc`i_/ Ō93˭"I+N_T.`pk`k.PHx3jg>LLj9+j$ˀN_gwI ٍ۔ZC obNn:9H2u3"ا^JJ8%Rb%0l\+dY@;5QPzaAif&XMӬ`͢5^gU$ӤÑwB2;sYwZ#W{N(X(6:ˋZ6N2& 6ldKֶ dlt xMF̨4'bCdCmvҊ6L*xJ*V@e!8}̈́1L 6Qdv@9.̻x0 uDC_E_%*uWu(/ EM oK'-~r2RDNX{W|5}[p {]"Dž@xwEgX'dft q/aN(E3fͬ$C~Ze]῭nlM*_:{6g}{f#+1Kf1*t4V{g&O#vZer#9SZh ]ʗs5<gY 9s&x#ƙ]8㻋}w1メN.~}5vߢm }mGP`OG5>!6ia tCb=,՜5MTN5f){MH֩~suAP8QobގθڙaAQP\Lonzah3KD}0NҘ`0EV!%y6?&KDT+VMܠr^5(G orWx'\]x8Pa޻>}ثR}z*fgG {l5q4HjvO4Jx% z'ILz??8LD y՛uo^1lDt@ica@+d "jbHy<ˆ^>ޚ !%}Ƅ\- ";RLGTc!K%"n#5t+ZbQ+rN\FmEHڥmSõta΃,pM9J801T"F* rNYbM"X&P7g92;)n Է b0]8!$Q&D('Agr'8F:1JJPx%0(^+#)vYJB)kr`0 ̷$`ϑhbň[\d𯴘Y Pk@a(aLT4Dٰ N <_qF^"QIP,4>Fj17pNs1hCLJ9%sS?e ŜhF{ޝNbPWViK/`r-T*dH >$8Vc\StycNޞ0 ԣTLn[q΂;9ggKP!Us, N$[zbht+o9 Q "$3QQ%F{lv \pV(A@ -Gg->6qf|~@ggKGi&}ia؅'HuuBzMۚײУ; Am`M%ӎ`"J!aܝtҧ&H:|$+/)ʙHitL*TTXTqy$" h1cPi+Sgl~ڄm$ +pe& xd4(-%"*<3/;n*DF)BDM"\)0+!l< E^)p@1Pp RTW`IB* gsr$)V8p^j!|o.0n8&5kyjar if-I`tozUӣ}-9)hmgV X+("GT J1uGf2 vp{48 Ϫ+aP&*\:ot&z6] 黓2$qQ/322>2*pۻ4")}SUkׯeJq9K)(\V0d#Cxq lY*9>R&[XSٺW4Og]۫Yypv2qyG ?YA[h;Q6jxwf3 Cf$XHnR? 3gD\G% 'wc/u#h"F[+%H٨*Is`yW0c|{"^wU2>eqn)ۡߟt?]T>;]Rj端F,K_tRpq/@H/O>{Cɿ?'ww[2 .Uz jo1{CCS0^54UЭ$xquK2)u>W! 0@ݾp[S&i%8brX#-ݑ[ܑ]l Ne{%=wau]o/ ?O𓺃F2gexPQC{n=wRk7fr43sG(5 hx2;1+ɭ1SԻ#&rQvH;$^ ae>*‹ A=pF͂/f"<4Cv;<}kC7FZ㉅x[ԟ.?I=c ge=^C'W%.J%5$.I@%,?K9^iv=m!*Dh6SH8,%& ]m2E:S{LkÖbbfBo/r25Bp Q^<igJlIǣÊG-怫P͔]r |w/:Äf\'Tj`9,bJa\ %6b ⫣( o+[ Sw\&߀h7B~`s;_H}74eUi߿pmvw v̎C/.|;YL:=V)n)sVA-m=7CJf/hlcN:Q<>6,n˅,=^S# s]E'Oܧ7%i~S2[C{B$R-mr ;A<8*\b2(JRj3jX$"﵌]ۘRͭod~t60+ q2ݜ(1^en(VS7n Lx4.-qGӼϲJn,U4ɵS+ʞe&pʃUA:+ZXΰYZj+ݮ#~Qҳyz~Fkn.X} c0ܻLW7t\&[:]MӹU fiFM9ߌm犷Rg'&|gnoN@⾛̢݊;I`pzMH6F`RGlƌ6>f=%9&؍A*F}1.[{sd\р1T)b+c#A29)wE% k"W[vs项A`0U9m3 7m"sBZpT v y A7 Eق}7OLΝ}=b?˷W NTn뮕(`ڻЙEс=JjqQM*t y 6+%)ll.ǣ_e0bQG"`"RSFDD b FрGbv]׾~9dɯrKzmO(M !̏4f<3Z֒,k%1qXQΏGhz_οm/-ontÞLy})}xs̽佟ݧc70-JREkCɦ!qX(,\KK9,'Fs+7^"}j L%&eNL%FPΦ!rZ[ȏY6U)FbM`M1Z%&"3D[5R[c6h>lʘZuvpSa9Lo[~\f\-FY "QBt){4-8KqJR)A?$Kڂ(7SPtj1*͉0 j]ju uݵ"4feKk]flfяLnհcZMR~Q>Ws.{)d"@logl߲(}AK!4F'dBm،0vnmU`~l뮓ɠ%xt` hx:PaV;D_ N's2bC((ч3Oђ[}J[2$} (AT"p)>/[H^(Wr46yAL`[!a 9p8GcsD%cA[ɋQ9L:'fGCsi{+rﱲ? t3Il'NbM0b:vKE{'ЪMQw1%={F2L4e_Cz{s^pġ 1YqYW_jVP-R4:Sն >!#TP%&]Śe+od.:n#B)] FBnޤlp 򟉏=blV,ǠFR@k?M s++2^E K$*V;Q?a>uNe|ASEr&jPNCdBQrZ%Nz$ꔟ$*BjQ4)љ!ҥKQhf`8LeoDw?π12I'H- ȮfbJQZAW1d%nt!Έغr{W8 swDj`{Bc[૱Mu\k1'>iIk=t # FGxIЯXKZﵭLClM ' KiMD|~w@v?sMR,ٵh@G L1087HSzwMgYO+`Y5)FֆszXl)&(Ly΄Fİ #ln$`NV 7kkA>~,/'~ӞqaG'bGlC-FX u͓&Nl`*NStp!IoCƖEh_Y i;vh쐤WL`+](Չ('CΩh~Y}=;XOV%(68R>T,ZZ :RYigz~QK%{+sg]&.FqY?_pEϟFv7iӫ33k;!kYVQϭϾgU~}}/&&Wd.׷ecf6+kVp\[B+\4UD|AJߊ+_Dtݍמܳ+Mbeÿ9`JznD ;O >P Tr(7Al&veXB`T2%L-( /6SkwNA\)!$-{Cl{mhF**\sa.\PY(.=)i iaGE?@@}LQOڻ>%67Rܗb[ yG_-U]f5G?1f)q:ˆ =75 2A1qʘ$D25^/F^zCo50r+KMH! Wm9VM1-bWhl@#aQ^{I?q¦NW/Z3s)>%O ;#tz.{gU#$%S;t9JvJbOAzTNK"{3Zl{u?^Me5.l#V/_թҠ\_"ﯰ=ENy_C}@o9ז4N.ľYKw$&y_:1{k>gma{\=qSnݴ7C_"w"so.[pf{Zj3]4E+wP.`rl)[VwF; SE{m<||B˳~=i-=t0~;m/M]|s?ߦJrÿd> 0O̝zA\}81P0(ǘώiI?{p;ݕoѴ<؟o r1uLozXb}Y>e,_fpiծ{!nsJxUk>`j<Ȉdg.1ǻ[rum }#iV|uؼߝLO$f}c&|U ݺ7mK9>da_e_'wC|VZ(oEwPd@n ʜS;g[p{_+Us@ǗSv=>v5? J c$z }o#( Svy6U)Z-u*1ZAk%̈(l-QAG@ SlZ͹:jeJ8e sO30w]6VXO s|^O●[QhퟶH,<^$f}4% <ńhN1?Q]v[ kj//U'dwԜ4kS6S[;Ϝ8! ɷKHլ W5$gӈ"z2-J?ko|6|nMigb}|Ѿ'>:}n<>}=Q甓*:u8R -Ԙꋫb0>WcA)M%-JkֆMk|")Jn6 Ғ%rK: U\`a?ȳ./eOW=${8Q(W,")N[i{{>t"KJonbQu&mf#c]>;~vqQ]gs  >jZTHB1s.) l"Qv!f I#&wƤcӘbhKFjkmǙ C@WSΎ&sHuru^yשȤ:3k@wY7^sin>,C݉9}O x*[)B%DW SHEmdPдV$(Ŋ xdD4뜩H̱@HBm!WVj4cep6,HR!_2G ^j SQY`nN!QC u@M= )tC dn-VJ3 sZMVF`V ;&IIk5 VK)G<yٸƴ,j 9FE=P-Dkı)J_Вo /-;=ƮZMc]At>Ӂ$tҁ ֚@YX}AR"y:,S) 6ɉ}8D-%>%Cr7d*!;R) Wbor%' 8vP2VzI 9p8GcsD%cA[^16ϡsbv4٩9+B7#x_>|++#և&$Vc?NU$8bJ6!z&fdЙhZ˾,GX_9}0-q5CLV\WbC=eU ,59B4V/,C_y#wqJq߰pk!M~0Ά 89ٻ{G +#er j.$km+GEsK؝`m4vv_h "x؆d;YGoё|m9C"b._EV:<4`5ISmٹ ToEvKZ Bרm55#hFg@L)ia*9h !(juZ|vq No+M1 &? a% :.JUi LQVXCVrNQd[ E{pUx=gD/"??ڴBtϧ G#u]q #iM(OI <ڮ-"3`i"`/NF5stثމXh;iu*|%Ldo1T!Vg!-' F_[ 'AS MrWIxjYrES4ĖBL 4Y2]WCMrL1T+zWbQj41` &/W*8&;DV)|_D;_BJU W 8>P9 ']GV m.H <{_j`|%[t*[j5%[j` {FЙkH*ǒ}]VZ4eR=TEPK|p'|ہ4"'o$`ÙJg])VՎK J 1j``ba_Mulcߐ/h̤- ̎TmrbOl;3vIbz^״Դ7aϠIGK :+uV;?AŋEkdWJ+l~rTޘZ]˸ ![9kVanN1p66&Oij>Bfd*|2ˍrx\O'&?W;C\dv4{v=a[[8 $jP[ `Oa;+:8c-4?9*L`4SoQe2az?Kj"NE&´*ӀJ}SR>"!>\zpǨ`=&_fzBJ.|XZ k9bv^ج3Q 2U:{ BZCNdl1:nJU"q 0p ΏRXxQY@5S2,"85<>_:>Q0ti_K4/@֟p)Y:y>ogڡ{xlke#d[;_W"}#Zʖ|%nCW (8K^û:Ó,4#|ɛdtDF/Rxf o(<F ԺgvԞFR@(BT8/Los|az0F\h=7{ײmw|[t'my"'ErjF<5*e-zΕ8xN7f۟G'_ܩq?Wۏ[afgzRNyo/v|sL~iE*l-vP#ST-}%V]Фl*XIu>V2*BmCwM0$:,yvxzjy{W;8h{mӷQ07l^ozz#];s&W2+- CvN+Uʚ%I$ ZH27䮂7#[:/m/a<Ǎi7B͏i0NN/Щ㋑_LWi^^{k4s{|=żY5tD޿b⅕bcX'YFDޫitQraaN 6?N޶k:l"?%+wg |o2Q`w[<_nbP^~nO93t77Pt'gn~:Ϸm`o/ð ꄭE&_UH^pY^m[76])6A_-^+ѹY<Ρ7:VЍ Rkk 2Frx8F]^K~9Fyq~oj6>Yf/ Z/%ٻ]bC<:>dE;͌}F#2)808gf?̚os?o঍llÄJT稵2iȬ)"lM>J2j-]{rN \.W?(# 0KC<@0cUߖjscE.Ej6*"l1H!󺕱HFFa4I MƪʘHEIP""f-2.&Ğj;JT6cc'Q4RgMڡʓImkkACO>nIRcTsNJ[WR8ʪ HGIQCNB4=X5Pf cUJH30f5lxZsJC.;ۮ=--!H=(. :$%LY:lSФ4IeHPQɔ* F@-$8}₁VsUEr}_ _j .XH! { [.`FLi?٧rvzՀVl4zZJcW'1X $ȦU,(U9r,#aD+(3F[T STBw $l:t(|O{A `Ke,/p:ǯ#x4_G HeǨ E. V .ŁHj%&|I@OZbsQ&8l iX蓴Rw']CR [AioeUX hn3,HؘG.`@WJ! QD^]..9 `a I%L D*1b!'|FbkV@Ql{ZS̡mť lguG( BbD)R*AVM67VeF}`yҒ+8цnU[Pv=x%VWf{ԩ\G, Il`@7\M;@\$U,V,F ee[FP5N3Fz c{DE)*(N@ 8eږGeEE8Qb^E $d&f@57]p` 1yP$rZb#MCa bW@ S!C$:3)[% qb e z ϤC@m VJ (){F v&#%\N>nXD5G]9ۡZbT]Q@ (belyD{S!_JA(D4PZ$6jed $zjqp*(_՗⼵ Bǁ Lo [ĀgmrnIRdڌY礁ZaBqrvP')&_0waFYb&.Lv<]-Ӷ%]PA63<]y lS}Ղ&aƋ-mV2z(9@:6W5 U@n2pLny'qbCE]E ƒS=H >-f戇pãlvs ^;%FFw #XT*|(a wBSn \CXQúZc4Bz>BV+*-W W@N;!I^gZykep ;b]E 1RŢ4Zk1Q9(X (1#d "eYq2"ƀ).uFʄHQ^/6“OWxpI "tDSMIAs-c|frmnN9ǡ.5 ʾC(U܇ciBۅep(﵃̌pa-t|m@[D]biO Í0"1>d4{=L_/ik 9$abR竈`h.ń12 3+M֓8nBDaCdnbɾ(.0JT,5hFNeg,,BW ^u#vދFJIiX#O:7L" ^k樖~~M^x`lRibvw{kMϺ*71D9 d1H879!4/-q[}ӟ]lKR4p-ǜ(i,-%˷?Ӹ~ܧqIz8(rz[һx8MଃϠ-* ͒A)ifDGe4l4XXGR3ٮlSmu9z2ʲt'4oǓT>~w̏`<]W(gk;ֹô8]RxQz8Mq ږwJFj_@G||8EW\*.wTՠ6p4--ѽz}~{ʚsiLq7G`ۼp Z]@W.z;C >[uq8v+V3 t`WzD2gtqErb^g^'?NO?!Omă1Xb+\ݠgAxBQ!9D2Sx);cٳ#KKt]c!1 (.$oܖ໔*Z8,kC1p؋VhXXI' "hEGuFom⎇`H!n=g)jq“UyC'+[˗*]D^~50AD^&2Le"/yD^&2Le"/yD^&2Le"/yD^&2Le"/yD^&2Le"/yD^&2L'K^ǵez[9bhҢl3A }h G/v6݃''rOpމt' ]7:tkv[%a勐=[x[*ζ@Ԫ2:Qkkb1{@M3a'򎔱puNoPBQ72ח2-iRfӍtTBJKoOnpWNxI_x?+^#F}e"ߣ&OSW+t\xdE#:}D|/A]lEuOpbKy{ wgxQp|}߼t~z]R*qV2u#s 2ORzI'<œecs>`1,y&zy(mmRM65t{*slfzc<6GvlΝ4f,OJb-kΛքT7gdaZu jg[j:w9k]5cMǸ7qnkv1ŵnqF}.16.f!qzWDž ?<WYrv9S-Gf737 `\D"2Ed."s\D"2Ed."s\D"2Ed."s\D"2Ed."s\D"2Ed."s\D"2EdK }HP:`2;gCʍ)s e\>hɬu8f0:(Es :/֟jºGr'"\Df8|h<|񺑡bH&+!r\ƤZ }lF1a7qnNX/{>x6nKX(L* l m]G!}K^}6Upv7ca*c3r#JP:fH#Q:At(D JQ:At(D JQ:At(D JQ:At(D JQ:At(D JQ:At(YJյ8 J6 +١P:J?vJGRt<JL 0 D@ D@ D@ D@ D@ D@ D@ D@ D@ D@ D@ D@ D@ D@ D@ D@ D@ D@ D@ D@ D@ D@ D@O'z[x5xۨn58scsbC w*%O7HqwHKؘI\LdG)ĥ'xk@ ks8yU\+\Ui{J4{z~u"y$:,īs M7y>xv&pbW/L ~s&gۯ/LƳ,3n8op;Gx|jKǹzp+};}`+};qX+};iJJJVړ˵: O:̺>QNS!Q>GhˍP*yތ0u w%k_*V?&g'{K xC"ݢ2\eB>%+V(U{L b>e~3FZo3)ƦE4JxUA_97?K!Wo.[߲FJi|mSJCPfq@.aسq ;q%sw BK]BQwx2WU`+\Uq>si%\U)!s͕qB'/f`OF7Ǧ+Gے;?mIp[۴71eѰu+ԯ$RI΋YI6 N-bݳɵ!em9[Ȏq)> g/j;dQ,>ZU %㳕s^qOO?#4xͼmLOY|x 6~ ٯbȯx _<:BmU=ڒR9~?j6~20%fmҍʖ7hZU3wUȝge[BvY[d۲-bV fܷƋ(*U}TM.Қ$^hDEN9EM*n 75S</«Bta LL"*ʋQ0gC41\^6T7E{B #7W?oW;ݶir9;u\V 'eRH K{Zf 띜üچTPvM y# ^T~ Jq8w=t 'O%L ;o\/\s^Fͣ"K_'7fi|UXN96!(Wm뜇hQ Cu9&׼xfم|WkbZG\p0}`q껯eAi}%e+I34GyF ]a˖Y{v_=7_jvcc\c}f;foWoΊ7S̳Fǹyvf }F'W'{Jw>LiItoR )-_UQw鴙OmGW_Wh1.׼n[2YsPHL#Mݗsr| 2WӃxd)Eサr97Q&Xs۝.ϓq,1st~rrkz 'P?{Vr /Iԗ0q&;dEfF_-)RCR_ՇQII-Oo#Z= ݨkob2mWd\8 f>A d3XHBF:,a,Fo"F1X4)qAH*"֝@:?"+co͔ cC[+i߭7i8ݍ%(xh&O;^ˍ%m*.+Ce@HNrMLetQ xJg8/a!2gd}]Ppۃ r#U/z/ʔ5.*RhZe c3D4T%yAEWsmM/+r imʴ!h>Oy,׹D Pbu~ޟ|:j?\4Tg#LgM81[:j>QZZD pUAHdǼ?Bm8 ?`f,]Ik{joPNۺ'"U J^ "8g\2qrNޗ F\?y}>I 5gfA.pౄ@V&2wW8MiBN⻲pܞr߮go0-r?.|HD3e}&GMJ9t>ăJQMpnaq۫⌶R^vpcL҃Ebf_[s?O~jp9?[~fehl1]톡y͙'w4Ζ@mo鼫"f6˯$ ǓVG Y,;ddd௓i{YrF+EdaV)gzlϷ8mRrqp'7.mKQ8D h&m⸬v]CմRpx<Ë3?_/?Xw︰ͻ_?{cꤳ ]ǣ: #:p~M/r`h[5--tQ/'v]!wԪKa@.e۴`69NtMlISv·ҕo0ۘf󀳢J2\<ݝw]c,d+2SW8NcΊWן'uGBbCM1eiv8RP %|dRIMȽIs[GViE1YHR{-s܉-!9M}EMy^*x8C(>:Ck8IQ`GB+,x1V 1C:P3_^]^t$cyAdZ,N~Fh'N[.>W[0`o2VRX]oӭ6_~0+Cc 8hќ C?Sw:9<9<9<ӹN[|Qzi*TcP$ !{WϨ!!u9qR o] xMk:ݾ7on@-فo{~B(%*~8Z%*u؃Q/>(7HT7ǴYeUlS 57/qiIFM&6^9˽i){w>}'=7JV̲:ORr&sYrNz6A0Zwt\WUIO?οac?-Gj!dzk%7PՁ\m>m /=U/wë\:i:tNڬgN7#QJ/Etg{*]bvf[!ld\Wkz|]xK5oܽɎVOws1'|؝k {*N(6T&\߭Q^ᯊzye~=IAno277N34{7 O;>\xDS֑kgg\9ڗ <<:f+JDKʹDeD)5"3:P)UJXh DT%ۑeam^u!71BZGa0YTwC$ xD]B96̢bpN9(SxaJ1i4E)10 HNPG_xQ;`ѡDŽLdG0/#ns[Fet_M6ahʎ[eSы%HR;`4Ϝ"&]RX^)i%s1 c/$!T3^NzS <"d%L$j<+"Й$\%Z[^+X6(HB;Ʃ1ɛa$ FK*`6[Dc(wU΁rv|9bXhNuZA%$OӋN%mHZ$,́HE||-/mS`%gҹD?zʉHSRR$8dZ Lxq)h{HsCjP+G=Oy3Tώ$$ @eQ̵"ϙ4*J5>,+*bZ$,v7D%^Y4DM>F{Ef1 hEmGo"B}+ JAzѳcjK§* C"]V)i)bu21cB E6%_84|{SEXcj+ҡ+KF?5JE+H5rHM^لJNo(tQ.6J )xqxS  I>lLWeJH=gYs,F pk Y!QIJ)p*Xu ȃRˑAH1{N 8$l)BV'7 6Tvպs@[ Ĩwſ nLitaq6&!y~ess #)'CGiߏiiɯ"1쒚Bm B3^j$"-,P4 Yjh&Z+>9^ 1 Itb6k1\ gqw-wJ2*Zf1-/1_ Q.\exal ݨo:5t9JB8(v jhpuʕ`t26Mr軷]~n90\;,P۲?*|DZ:']h6SԆQj`mڹ X-WtX&O_Kۍ4*GҺ>7^Rq[S eH)Ci}[[_/ ?p/)ܽzx#GW77۞1԰J"b$5Wk@* "4܀R qt)qPOg9,5Nۨ+E^ pa ! H*ĬTC*Ņ7XV*` Y(X;[9E YTĨ֝5䪙6/YSXI!"-7G1_b;ݲ~DhY&y2+AU m\ƣȹY=$Xsy{,3F{)3Y:D 0QMY0̑,fY[~ *<02`P.,q\8-<V"0g*j904\L'>I;tqىnvumn٢F´DÆyv-|T͘0h"^HȹLĘd*G$Tft\`Wtk<(ߢS[r@f!TԒP"Rf8qɓn/}էەo;X=|f6 "2IO BGmׅͬħeƨx̦ɜcV&$h|&Hz'=1ek׍@shSRXeLIՔ@VK۟qkiyFBm^0OEȿ9y9 X>$b dDEf)9̉Lp\z.a/ŞwdX=RkI>az c"K}D6{`qyQH[L3Mw` 戙k=io9/;yl;LfD,yud>dhYڶdӀe&EVX.(qCvYI(J1,,VJ\7g F&NҾj$UʨurPFWl&gͳBX<}"wXU3QVio ,e\y| &## ΄2$a6jF1ޤl\zoXfJ{B}Y))&zBڬoW12#XP1g@{g+Gᅽm*{ fd\VVÛMϦ9u ? p^ LQk %Q!`V0/2&.DC'dli,$Aa iU3P@ T&Gb9AQND{@ Ҙ>(2X'DcZaR ,qSǔ&KFb"$yTҒ+bJ)9b:2eJ˔vvD5ёl: 'Z+g'l_LuTak!EMqt< mĔƪ5_*nq]~_D1<,e,!]Znu OXTvyRT.ϐv;„tW(w$l1qV{tHbaJsܔ~\]qO dJw=;6%qu}Y'[P-Z-iYcϢ#z!QVO5Gb}B7[p)%TS*EDLݠI6S4WpSbnUdvazIg2 @;y~Oɏ'u_ kmly0u鬭Z<2.{%㴴ڪr | `xYL7[I7X{&CD;}KI RDpCJ,s$#}M3\&xÑ1Ku r|JN롹KH"-D AitbP}-[giJΔy ef̄xyIAWLX(( C&Sǒ^wcI[M@jK[l؛ ݶht ((c%2&.w9utd[rH=d}w/\\) B},qNj:TMohbiV$M5E݁EcPǯGM+eZ/U߁IМݎ&nv6DI*}5k8-~R6<))jOx{2`}tТ6X!v.11y^'ehB nD8?x\m>駷qH?rK+6m=땗4:?=/\qBqԙ%a_O|58V^ث>;xL?#b~'f q}y\'a3px@j3h5_h[፠MzH]T DK30ΪM8[GU{ދƓT> opAL : m^ [oP hġlAohwJ(vҢ d4!tP|UĤ)du)|;\u_j,+Il^#C*Lΰ.®\m;JP 3dWk;î\鮰-zv8gȮ bW `;`W]aW-ElU̮!҂qv]aWjWUl*AIQfWO]C֙To!6Cf)aarh:VG8CN-Oۯ/_C^4}%Q ~-ԏ,U`EU5jd̦o o &]j]6lNS/_R˨bjFގ\t;3Jx>X0x:^_ m108٬kLާg#(_𭉎9nV]F}F ~{hw\Xh `wfMj+-@ˉ6AHmnK$!h0;î|gUVmgW ]=#v%IS \д(߽/ŭt%&10DZ29yR"eλ^}GDCX424"^Rg ᠙Lbb,fF̔ d2ʍnF(PW\ it>2)WȼҒ_{cC_kZ\x5IxLMQZRHc" ouBU_5PH[ EĀCi :@y܀,0GΤR&b8ì#P#ts09'L s09'L sp27Bo3t{FwFv7UqRYJi# 1͵ӈabHlєjz(G neZGEd@ kZ U@zf2Pn[Խ$MoBuD!xnMkEۛuvĞ7,<ôFh\Z P\Z6zSJJe,ԉR";Gv|z ]^]R?`r;n90sP &Ff GHTR,.8i2"Q " HXG"\JlB O1j(,ȖA@kYςloy-jηH%Džąǃ=y_}s}I^h>t{1-Z4s|M7(Yk!=A\2xHv$0"0QA1bH $7$ZΧIBP[q]:R@4:&@SAN+!L^2q*ݖ XE3\[|u$pn,DDY (2VBD`#.2,W>ښ!$Q26I$+KHMl< E^)P@14Rt /|30c~~{9!ٗ+\u\`kS #hs:,"¬Q"1Lpq_RD S~?L6i{wҖT*}G+:%e(bKdMSb(8F"?IA)nL3`'a˦ E-6`M )`\^!wd:N( 46>&_I!_Yqj*F` 0{~kګ8`QIaYnZeա}UCa [HRu\0vQWOfFZ~KT?g/9O̟Q>}qK9n EF?gUHha1:zuk{bN!ԉ"fy|Ba4x`0a1E`jp{pdW]c^ɲZIsX>O8)X$Ÿ6_(F| 8U~,~ȿL]r;k]`*x2şi+Jx ߾RƓE74qPabVop@$ǯ_z:}~o}sǘxy e\ _6 /^:c]CmTuu W9~y_&yW,D q7A.\eȺశRG(Mso9}YfKT9( ۓV=Z=*ƾKdAF=iIM4Osc|J3Z 6EQ^k`7$Ghh2;1+ɭ1Y{hH ˊԿmD4k]2N`kabAP4A8#gA3>H ͡|!V&NYM{w$ףƙԓSKQռ94su J0LoW rplW!rFBB RbJ_A6;ot=9{0sPaK13lK ɃyM\hBA:OEM~n OK;x`Q^W>yh){RbxˁflFAcrz 6Sݙ;hFS:..WyHϚ|:67ִwbe}@8&͍3SqV F<%40t}0cVei*u TOcҿgPBG%J(I (4 Nc: 'W2j=A^E +7䈎E[B7z`4[0 ^س86 w rai<,KūO@}v['7 g}W"FmQFL-. 1Y^},yU"j8!ϒ* f |M6N$|ϣzsr1xr?@_'a3pCWArfP ucE@ K6&Dg.|<W%ދF%_xv7 bPv6U/ ua(WZ^YPvb74|\ŻQQu^ԍ6%mDEޙ'i|廷Em؊lݚ Xx>Y ߃݈+ݧ^cTu2BAdÓᬏ4`ʔ(]2I}%# *kUYz:bKPJ`* FSKT(eH9Ad^j AHRV2LwZ D佖thnDUΚ5o~5Y.[0tqzvst:鴝 k4xC:B Fͤ7 M7MOgUMRkQzc…K@Nޓ+OMnEM u1giRcRDtbJ:q*^iriږj^ޕu$M0yrG ȇ$ 0>eeR!i[xIj:~U]]5y&s󄶻)EܳGw8:rOיgesXb[^ڹkW17)~sPb[J@}`'o`@qMХb:PB|23<qPn”/[cx Q;Agm2O#c( W@k#9A]\E-rvC ٙ*ȎeWݔrNɥoMS(> HR#0 QRdH;N 9V2(e%6jƗa$,9cg᳒QF5^`BuۄKHi) NYƲuOs:Fbm%nFb/Pdf"fMX}̡,g(gW>&=Jf[iQ."EmHy H,̞ h"PS>=+z1L"FIU4E%^B%ZSCTjUZazI *M*e$ c6! V"9fQSxF2⨢8,v-x&eCUܩTh|B2ihAYa+\K= pgC`ED!JCRxzr,WE>UY[]7YŨ4񤳖qZ`f.11KO4\(K0Y[Ee}. C4WHƩt%#jkP4D_ec+!mve͹rXPYεhTҪhdnA PĽsҗ-2L;N:WI (c<71 筯 ӡBz&5wo+,?q=-]7ddyEx!K[-G8@H3d(^!M1 ]U\[LW5@FH&$,6dl O($=$إ?Z=D㰵]sv8}oŋ.aL!})|Gƶm;1!.ͦ2z}5߼K}鯏WU> J>4y)W ڀ`piVѳmK?[pKоڊG6νvmj77WuK4݆ -,bF{Mi @*ˋ4m>G5?I`f7P7bnJ35cQGmdS ꏫGY=ډgt@ת֡ZOCM!XPSs*j {Be}ء~U~8̺|+ͧx17T$`#1F\Ja_.5Gڥ *%pJ9`uy`I ,E011ZfbHD"Jΐ5 9Hhd31;4tsGC?^5dDK=Yo,|ƭַZ#>GL}хwdbMbns^~U)R9&plo\9qŬgv,Y*Ιy1:o̡qWmΤ(>_Tp(s lyj l|&nB {w(w]üm٘D?cR-{KrOOKy߻%1Sў5H[iE6IS%.JGUYo9=pr*9@"d.s.ץyX=RKI>=aO2?siR@~9\VRk=CrLp9I'!]N$ن8lfP0$O\Pd Ie+o_19xHInJZ \{1̆'kӡhŸ=\LU6:EdE`F颽tK{lK ,Z]~wi]\TM%7,X{6\6L+w`YHp9Q+YH>>#H0BhKq⪚Q[c3[ [TQ{.E38Z:ĊP[3V#C{QAta58]h{]pIQchncVwqF`Ize쮌匕NN\'E,>.Mm 5A5dlM/ivB<ӗ4}hUTFGsYK,IiwJtd%+* Qd Cr ;Z^T!^`amma F\!1,K3h|vjʂ]gؓO0-3o; !&Xg}vxWb0lP@ħ QcύZ[RtQ 9&#MN?|JCpq9k13߷2̊Xp{r$c*\xlK~U P\bePЮ¶2OG-fhA픟V|pox~jՅ9ku !tpA@sv/z烑g !z}_%9`5dV,v֮wvmB9l v9!y)Y}-@hW `+iĬ.ϵ^=_Gp/dp]6ܦE[s,f3  7? >$JQb]ÔH4nr/;l Ҭ(_[R+JjMGRQs OծHk5g61jLv=z^'5݉%wqj5&\+vG7MuvS.3N7lr},_ɹ*A_~"J(K`W}쥓] N(2YF8$+o͖oSv5YR:/xY;dݬ[kwr:5o]]Ux*FѴ ,Yy'ek<{Wy~9{vbKYօme¾{>{-y+Qj~k -tWZ"uo8ZnQ l^i ]l6wA|uAT/^޴y/{>_!f,MOWT2=R##\iٞv]m3lmp>7C]WΩyJqP"X1u2qB.ק-JyqB%@}qP\ yBɨB$j:vuU^]DuEZSBWD8SK]rɨB-*T*իgxǥg[gejqokiAsVyËO+ۣ^Nn#0`\s&o? ƣHyE3F57"8Wx*P+BJ 'Z0Qr1;c~s[@{2Kn2->t!BD46Bt+e:'IPb"Ι+2Z@Ƶ\w[';M4<,hwi1%+5*> Fe'/SezLm4ɭgכa?{WƑ fLKuEFY; uZ(R!)bꃗ$E$%v~UWpP{Dd.n0rϖu,ĭ%fNL;T<-󣭟<'B6TZh2cRuل۳ Gf~ԣ,Rza\ҁ|Gbԧ N-}K\\)SD{VOGqb01 *8|/K)%3^&GàA^_ion_Qd<)VP]&S1тFQ4`pHV|}}7mKč7Ӧfys?HsVmюi{mӯT9FYet|s|%T8 ydKI VDpCK,s$#9&)AVR-M ,urX Zl&1CiQaQû ^w3;sŲ8^:xf u<񀖗S I諾rJMQ*57li1L6K<]7z<(&#w;&)rIu|[4y(V[vRKl09- -AƒEiCJ<hgbV&lTiRH'`Lsmu韛ӈabHh 5YAJqT p+:2,"05`rX+"D4m٦Jq۬ l#!v:|?Ӻ mأ?ΏKFg AA@k qF ݤ(NzoO@C =^3^?@B %Z|gR&<e&U,$-ף0Se+FII$\ b$v, R W 0܁(f hX,s:p?j~ez07~'UXz5cLEl-ZbjE&pc3C_Qǟ#22Fi npwiؐTx")9`p&ɘfi%GM:^2X(&X~, Lu["zu$pn ,DDYL#cA!,J !0"'CRM0ҡ'#e}%%o A |kv⒲vJ.?ܳZ$5NӉ0>C X'(1wo̤K<-ˉtU`zuz{e$efX1zSDNoG)(ԝ$\d B4 tQ?kUE-6`MCOhg!}wQHTpU>۝dUݘRԻ;DoޜMo?*(sx;!.b³+.AJU ・5խ{Utdu+|;?8_g'C}68. 5/p0|jI :GnCD0чᨔQi& FOcM#h"Fm[+%\:Id40h\N@~^]:ç7u\-f{?~~O\r2ot5%`+*JD<=|{4_ Tҵ*RV{s L_o?[뻋uq]'x/RV)胧σ&@x0k2tU˸:[Lڗ"D WzqdKΟmGE!V1֯\~rG(5 v7$h`X2;1+ɭ1Qܧ#$,JVrL'$^GNqZq \/* E$ 룊.\BUI63a$u{Ue*dBƋm7+Nk0`2:}rn*DhPH8(,% ]!8spg X 06 {镓!y8  M"H牵H;ccu'dxt\^:y1 y41!W_x4J̀qYs,EL)lрk1F,}ux\E:¼5zxwՔP-M/ƫ#PbAuo^Ⱦєjݾ)EfQ]IƙbE}V9bѵ`YpIc c`HWQFuLWZM]7\{lJ舢$^i8:i"iLrɕdžzy:+PRNq,N; 07l hApoR0#4'(xAcfm:(1=|\XX7>e O6}zkG|(CeSE՗ )"j.>KUVا7dk'}/?ߞTͩۓwf6Vᣟh|؄Op/ 'pt)ϸa]_$mk%Y: @fFYPLKJ-4%2;ySÓ >(|ფ6R³uZ7]max?&϶/s(uoݴz7QyfzFkn/S4h{9⮾co^>]UosfiwLvgϐ,qr_((\?kC⾛{3?{MrUV0#*(`FB_ PdgKGf]vѩ[ #LH]@n0P0vY6 YHecc.*b&>\@4;Y}^y-v;𥌷7%{.ð7vI%~C[t-6Uwj84l VJs_޼LYL-*?Sԙ<?EePvH념U}㇓ʳ0R]Hţc ƹH%uAG8jtMȱބh#fF9eR:)1)'J\n,CmQz"}$+N)2@Y.$p2ua3I 9z >:B#Ck.A*+?+~IA1]0maccR&ts籆eR"-#RoaR #A!:g.Bs@Z*+$XKG"ŀrKM-a$EV bD3lgl+nƣ}v3Փxr4jz-;_efR954|LkT9e9J$BX̩ JbA{`r#D.$rӘG,XAc,6 metۨQ1,%! Rȃp198tsZ qʼnm:gϾZ@6)*v,_Ȳ $TܜW(Z>TS. P:Sdвh8U*R58őGgw#h&뛫=,D4NF%B),+B@RDeQ\ }Hڂuᐼ@E !I l.-yA)ıH1WOpn k:ہd`f4#l^y/`(@^: MrB+#ZP0Mh%j]|uZBSSAY_q t $FŌ@Iؖhq{5uXpp.uXKU , XV3"|lH%I7{,IN,?ڒv,m M5_=HVѫ*\ tʏ}?dxtFxi^U*Z &"dfrxqݮ3]z6Z#F*{e@>oBhВ1%J'SW;#gX0,;݉eGeS5llƍ3j}yR8S,sP Q3 2ȝ9zd`!ۙVdLA''`Gx>KIx.n"&NIP*h9,*MتI20=YeVZ"Х|y*81 pFӁ2Gr{`Jv H=DML0~R'} ^ ]$!ApDzuHBd$A1{H5((p!Yl4w&Fz#mTg"|Ȯ';6 IΛYƨd< 2WB 3lLp$H4]P=ua$Lv" Ŋ VAdo$=3 ^JXׂol,.sj2׌_GGVME&#TRY@"..t5^Y(v`CȆʺʜ Q$u!YP}MvpKST#'-+Uۙ@Þ{؃k[daMć:{sxfN%۔G>刉b:sȑG*m6%1sgLI$))̿?I^Fhr%RsXq<=dya!$唊"L`6L[\Pdo_19xHInI(reZ3wE{NuThB3>qںi+&Ӌo 0__JAwy޺'n.72-Yr,*oq߰Etj\fk8ɷRp㒨"O&jR1pDdZgQZ&j(KzSeGN'uz~{c` /y(Pl4·ZbQfgg9CfU'F?~2',;1WBwu2#ͿCH;mK Gm B-"-TJ3Զ/\=*jq,pU5Jepʒ}#4P^@Y;L_ svV٨U`U)@,dcDϝ5&`J@$2(5\:MD0Ԏiqa5<QeyW|<<P;^|Wk2A=FC5 5EOtB~L1p [H%9:'J}?yg$wq}9Gfdh{ppå;6agime鐎!K &!:2LEFd:g.1C4hW5r@& :wU1׷bYWWJfֳow0FlM t˗@>GD|YN_'7ɏg !*PjkSn|LLum\X% $n"S2e}FCrڿi0 |ns4/-ڵ鱤<(0kXM/7)7\cbD2< fmY6j B(P ϵS CBn亀wZr->ndZ"!Eݏ)NݭKx_eY/#K- f/'<^B[# V/Y(.CL嫊^]J7u/ڵ< eAdGڶ +24Q&S(wH(w-#mrZ%euHX$̌w%Ռ9;T'cbfąF}/qAQNZo~:56ݚMU]mᚧPzo)L!ؚ@Jn%'CVCP\˭dx>9i-Qׇ t2}A]ѲʍӴX~|Pa@;/*Wy}N|.m}gm#H]~ XtPB!Zi1+J2}zd`Ӭ׹c 4u5Q=h`f#[-nnlF/$ B;DBP`*Yü)hԙc9&o̝L5w)q!X4qXqƗs*&!B-Ѳ) rGd*R44}RۛCr2 J2^]6tlHRn%l2~5*RhfQ~GqjZ'oDJD42]iIM ̻4*0;#//K f6燛!:TWy65gĿfe O!g'gT+f]OVͽF^_sK%ȑ +xA2 Bw qs~5;N\L)T0ŧt18s54D3jy^D/_>{~b+#-XRZQpnpQ$49銏Bå)Hi=X4/aFXs}csͺqvxMgm8Σ܂۵@yG7^M(_!c-=)=m݈neCnfYޑ 4E3Xfhxte] vwA6VMżY짥%'Ӻb+JD?"C9ı& qv88bcP%0,08+Dehq]Tqc8A}$!￾͏?-߽_~>pa>ۛ~~MO^)ŭ$ؽI^7ow e`p;u3krK/juok1dCfŅ~47r#ٿnˍ@sYigb@VSmQm;b&HJؐHJY,fU!^&R_k7H6zlx.jj84?iÛ20_kؿDgPAÊ 6}m2>4i7MJRIUQ8eIni5՜$1D2l cf68u/Aaߎ)Cu;v %2Uy0@DXbQ;+Ȣ$.Fy|t-q"Q6&ܩcNw)jE“U"$\_[u[#p:LB|R{ݣ\~=u*բ)Rr0$@i/ٱGŊ-G!)B*왲/ s9Į:%rLugc`qUXRh^7[VoNwzTy(k$+>RAx-0TYS.Ujt.λdowPkrK4lErŮj=o0 *wo9,/%i&I=7u(CtFʼQ}=wtz)\;-TPU)"C>V-)ʚږ5t_KA 0[[|JC:0eBt9d"Ar h,E.X ;S;nC2c"{zezdW>޼ʭ9BySipwIy6'{]%1#_=_O_w x t5CHUb FVii y/,D&d&_AM2:R̎mV K+MR%j{8 &1tUW0-ZS &Ouٱiͯ _~[˖Pu{ES['">Tl-*14YxS_|s)-*ȍ-{u؈i!8fYv8(n:kwr[i7|wpa3y<\^\l8ͧE<~115[\|uˉ9Oݫ+ϧ̳r okλ?^_~Lc=Ђm WtX@j[=?:V*P}L`vl:.# )Rz5:+hcBUIIpX{QPCm=~ QF60:4LԍրdX-'#<4h\SdYw_Gہ~ºrÀARH?塑~kSD{MI&9 l*hVfUe,k=A-a)`;}I+cCz>%K![ l[  YYoBVWe=ǎ.B1cT;Ջ?¶n? W/z!|Ӂf|%\.5рAur9cDߘ{zmI5J %Ϯ؄PĈ-gKro{D(9t59*Ɣ TAX(;o@..N]/g#~НYOkZԶII^nR Bf+HGFW, 1h4#Dk D2VJI&sa5gyq11A U>>.d,izrȃfyV/c^<hj[]59xze/z>tN$2L,/qV%3ds0h1+0ZHQ|u>IcIȏ5d 2Ch]YVA:F6m;nwIdX b㭂3ӡHt%b-SO[1qKn[wt!z-&AĚr0?5+ɞVrS0{4o/=|cZn҇ю$GH^]7gPwՇɗ|Z_&Kes?7' kBvQfcxѺӽXV趌2"S9=IcM?Jk?y&qN0/oӏӋhiW= wߑB{:Ǭu|ps^6}ILyߎswiZ<-\jeG6N#: J{o@/ߗO'P`IZ+gG:xԣ[K_ȲFemv#z@'FV'ۑiiiտR t)y1lXV-B~e@aHܤg%ȲFYmlE)wGG$0gލ},-;rQԱYϿpjBP^G,)$Ia_f桤<#g{n,0iyDZ\QZ,ۿ~#(ĒDe6JU_Dơ}OKwsqt{Wt<|xEv'3LhYa^yvfŽ\SB*d\:**1$L Dtl^y:J"uH~&^֙zsώ-]Nݻ΋v63j C6:ozcMDĭnݮn7w9ltf{9v-n9ڽM;vyfz77woaY[:΁ǣf9i~ϪΪV-x^d3_~7.=<زdbEGiD.hWNyk Et 9WH.g*p dЩAؘAiVAJvqZEIǀc8`3I1;%{N tv41, 3g:Ajh坟^5ǻ!¿'k^˝5.wI+e)J>Iҁ,I-l @un9$<M^Oj~{u^,x(YXUv(][FA]k<:A2r+ˏ-A1~MoJCg;̺ "N KrBbDTIymQ."rMAKчj2mz7=Ѵ #)EO Є> dCČ &imIzFꖑpwx'4f!gťYBĒef!F& ~CL3_M.'ujJ{dMC>deJ qY@0y~>\VOWi;7dϜW*4e2S!'g"dK?J~8:8YQLVh<}r5Z.MX%qgi.zJS ihڼ\铟oNg_]g'#q|> Fe8p:[ovPĤ64;яZ5#٩ʑY=aa];2cF42wb=ǣFOn6-N_G/rݨ֪@uR/2h#u`eTw=aUEhʭr\KNi)8 HGqFGWܮ7,3GGw7s^kt6ߌ8[ֹ +Cn1o哼@7neVA-٨/j Z:FGi^ RYe_Z}5tZ#l8HuHa}$p`^A[dNzPVⷾ^-uJanU 7]#W#.bJ!JsGl0џ7T264ژ̐F$=f{u!ڹd&e]`0!jAF((ǘEF1Ø;PƀT(mq b6I<""ctW:*]GWu6,;[vbL0&M5nC#{+]CW^ tG,\rfn~Jʲd@Y D!0P?B=<Ԙs6(ވ(+;/sd (`T ,nH ;RcY0i)F+-m ZR@EYb&;Mg|1r7>>|Mnw߻ڡ)7tf)t%<\zC LFa,4Z8k#gqf&cY%LIʒ딃S"F^XRߒ7)`%ĂIbIiU,2 1w6g}|\kuhϾZn*Z֍]WhSFE3duXH6XPYX(ZeUwһTGz7WPU5="[BrC(t)SH8[0 2ə\G֎j'-A$T|๱|RQ20x;@mhe ]ɮHγ_@5＀3 "@F(8d=:UFhfix@!H))L:dc0zYcS`{FD0!8%#Hږ=t{~PH0&@>j(#2zX+) J5)b2e JM/AI&(TGd^EzsYo{>;Pd,FeՠV' ni U8Pn3l Pכ7'H^Fؔd1$rhm$%WYgّ61Wuヹ\ktV5hEg%u,[Ǡl1%Ҕ2FN:F% F'JMCΠVzR'~0\}PtR(H571eɛy蓙L M|sJ|EJ -y=}9[gJd dØ*Fx,ldGKzB尤w4e--եn&v59ڶpsjԮ)til)M{dRy wD\pJ㑑gBłu05ko">YyV(GkJJH1:I}Ԗx9%.dF֢l+Xdd^eጢl P߬Nmߍ]^p JoO.O#H#%uk&8hf)N9"fs"$PӐ? JؤB:`^es) Ӂi]I;|.r6Kl;9Xvձ֙v`׮xlhf< ǃҒZ!H YxH2AN$Wgl\^yY,2 Cv4H5> 'H )% >rA,+>e=|Hh$bD)gum eIXYlխ [`tHTQ+P5"puLaV^2:0(E)r.fIrg6Fft^:]",q';x"kw.-7wIV+Gz9C:f/ν@}Sų a?j>7RK-c1s)rX:)fdrHxʠ:w~[C,=ڻuaTїl :#?1RE\y )$<3cĜE NVFhdf-,UV́pzL_7}_Iy%]y :ʠ9-o$Wwr0+h!յB8Icx%5뤅lJpucQ%baW盓:Shj'ތ?D\1\ GwO"!w\oFibwOR;E#OΙ'a|QQ&(%̊*=@ owN^ P_NBxpo6M`{Zԓ;|>//VlbRtdhrS !MV|zG9ٖȦ؍zדie3,ڲݐOK+cOo{#ɲo:9J2{s3"LD5bi1g;Ja8odA;-(p7|s[Z6nkA$fhkjyʩ w3;pdYwWXS<3S-YDS˒V{P;In"Q*57lB4 2!O7VetRx'Vm(_.؟U}y3:~\k&f b$4c{ E:kwh6ostGO ֙M=MKZ&iyyogumv|52QubR}sf:nx u"":׮諙4? ?w'cL'F<ԵQQz9u 1y-p*,͚aol"XT,5ٗ h)a1wM\lXcB;OHhy]oɿNA9^M8\_4e jb"e΃ ȎW>Y?ɑN.ʔ(]2I}%# _*LL=z"TI (#6S1O-%VS5NxFH91yV*S x$ )[X1c2b=6MVHKDf-H5eI^>T_u&OgBu:_1{"גV'kfK5>2࿋ 㭏Ûwe:cpvf lwnz|8G繖h8\`fo淪M:&\ s[{{ʊް4?m.F77/oG3R:@DZTJ~1u*R:@-FT$* T|u*B[S8CZX+)%2RLy_ZD)P`z+$nXǽ}{v WEpej{a" .h bhg߃v\{TRBV wG9bx[UF$IUlqQb5_=q-j+Ӕ""{dLqbX1SDNbNxlt<!xulM+aP&vC!%&Ϊ#YZEaPc,~C/L1«pW\J43f鯒D_;6J Jq9KիNV`ib³G܀FgPu Դ.T>VuwW7\4΃<_?z_sKmRx&oB-#nʑ_=U0}aG`DJF YL|8M'z|?f7YpZ?!Fm{VJP!u]ÔG߇#E:{Oq1fWumT6fߋov.V&Q,C:3_W 1 tPTaQG ۅ]_>}ϏߞïϾ&ޞ}|-¿{괕d~y?C퇦(0th6Cg}f·W|q<.~\w+CKWIy|Hhqwr.caFIU&a8u7:{8gmc_dy4z&_n֏"5P1~F3Z 6EQ^knH9΂CS`x8$tFsL%aYc!Ey堌`t`εb AP4A`A >H ͡*!Q%Oζ&vLLYMVlO9:z!t顓c ع=TSx᥼Zy4*DhTH8(-% ]`Sʷƞ;{|kkÖbbfB᳗^9P!Є(/tX36v&Ԛ8.+3"0KI+7z1ZǼTj`9,bJa\ %6b S'Q$c,۠?Nܫi *M- 6'P\@u2v|7oFTg^Pre}KJ'1CၱXja6(SZ8<]);3pkP FQ`TgͧEmJwt^}n~\D*w'|gnLv.$Wق;_,5FػFr$+vU ނ4P؇Ncg ^<*`},Yi+ev&33qx FJe(t}f"h3κ2_s)UVƲzeWFmcZ;bZ3ei丄|9AT6lR(NAja[Vyb'R3A%bPJpW2SK}3(AjVKBP4,k. c[nFj}]6hsGSsP h)C|? sL)m'-% yƌK~g~>30DDV1,faaFTNET,:25'eO.gi)L2er" j$OΈ&2f\f2&Άޅd>zJ'OG߾vG]CcBe{ ^?!w`̤:Qp Hcq$#1T%jx^:pI 8}M&'P$pq-8[<q}Wȹ:n+uOõگht:"xz(Éaw3ꐘSh|C_I]7dV)2 fWjȤ貊[U!KJ]<[WZDi1 % +37,}Q")YllD66O&lM^l#u,ɠz kէbXmOg$/ ,}j}BSDD6ѷE7-^:.LNz 6gLLLNȲ@. ˓ BRgM"YYΑ'iy%IvBRM `11;h|9i no*~\1 rBLAh+ QaYf'9pJ #(,Kɋ,vCg޳䌎K s"AV"Ч@N2F[#F. a ,+o&aY!֒lRBO9+*Y5q6z3O/FQhVXMJΝV2 E `Q#a+ΓeĂك:`QL zzFF U4E%j/Az/<ňN@prժFq󓚟O*J% $Cjײ;ibwI'/w*Nlq+{eÎ 74,`,w9 rJjL> >[)@J,hjuUds{f 11 =$a 0fyt)5h$62Vg=kYNf Mg+ʻ9˓}g&0# :w|#XPL1 Mφρc!|;<|J`b4яm2K!ˊ!GŵpCϦqz?=ƗCR >ϺFxI.̝ v^ye=&# 8zmMzѱp+k#ݾ>DT [P!R0EYHḦl:gP2he֠Q}h-FTWw믟 أO4!F:R2V~1!nLn_Mc]{wêO6M?Og7joEÔWj"EHƫN>% cҭk-tuJ]Y^!f9:_<^ T辬3ߨގLTov>nl?{ϧFz̗ \蹄sɿ:gB蔐YF117Zlܼ /ŜSOʟ.wdz )ܛ7EC?B'b">ɢwiy{;h^]'?[kk7\_34XetQ_}Pn53.ݏ6ݗU_TUj07MQ `7b29+nWdJk8YdA$kM5fE>- afV̄ӢoSaDt楦+gz].].{,l swydzn®ذQQm t1씡!5֦GaaGYiqŦF]tPnL(wg !}v[.:N+vw[<_D2B6?Ezt[t̂30 7H4 '7Z?|wj6]cE\k蠡@|ͨMx;RMrVŚ][~xF=O^KI#0E/f$+OL Y?(j‡Hf:0ӁB}r=ҹ,thz;*|uHkWe8gGCP ]=,o~8tuyvDלͧ%&d骟]yI&؛Pۻu޿6N.fg+QOsd_[zV׃"1{WE\HZH +4> |ઈ+Ҏ[+c6GpE#𽁫".gH+H)E7Whb`oઈa_H*RW䆟^c lōm^dT'XZox^:(M~&i<3\: eU3hTFݸzF Q֫dϾ^-7-lߑ,Y80_&iq!k:G^93?kYi(pCm,>_CK]VxS3&귟~Q̓Tj,Q`;fnY8.!2dc~z@b \qA \i7zpŭ  \qƌ'npJpn˧RoWr2L헿:)){F ԩ[ ʛәܩ3NnbFKmF5rB0Ss/ԓP0v!V:??6헵foPr}"2@·;f\ 1$ (GLzUɱC:AБ{&0>U%# f3e{R)ފ߭K+Ҋb.K+Ҋl|GPZ1V̥so1Vue^(8%D̼9sfy ydcfl|VRbTWm5w_$ӂx:O* IZe99,r g:ш>Fy&\8( &ePs_qfh8 U55z.gҋwAS?7Uct`AT6'1){-q~5䕂DկU]yF?L*S/[ej6эrgŨ5h:Z֨5h:Z//ǎ'F :QӜ6!fȹ&Vd3x;.8qS [FQ)lT F_[F}Oy@ LyDj5ẗ́O' 6|$oL{t:TMe 蓊ƓjpyؑjYHl$[+"qHxkQXiA8TtIh9:΢ s8N@\C Dj_mdTF"jb ހ[u6^h`&m">r> Yn<gIpnOܚl} N7|=kf@a!!(4j5W=ڷGAIrNe&ZVSb=mcxH%хT 9k ,2,7׮@Vq!hn$~FeA:BD[q15ʶ3V-O^h59LLӋI*+,//M_jWV<]ъr|!#ZM2~Ͷh!傕/!Q2Ų@`k`Y蘫V^E~ $¯ubHGӆPo9%]y8_LjM7Ԗm% u(,P<4{o%wr_j%]9)sƺiC0Sr,KP#"$ X\K(U%k*ŚT &4ǚ:jL%:D,dMY5fD @=dgאzX6m~&>;*{rˌKq.Z,^8O8}ܟqgdKҳ/[JzbOO>.;zzUe#W`R-_6 v2|r:iT^Lս<Cc3.=tij>.r-K$V8lSQ'y//˞R4-wJÚz7o^:<;#ϟ_?V|Ook /弤F *A0迾iA*Z[ h)kMԳ+rmW@S$BKj\S{nw[y`RoXQV۞;0i[w-s oGU[zv]R(xoW;scyu&;l-Y(:<Vb!ke|*wgdrF>gd죝q<ڤ#h:{S(mSd{̾M:sh[|t6{aXtW4Y0n>~s.S{D2@V W%:YeuLk\&V|f`*1jF2>ƻ6 [f/_Sp#ey`(X]vL8.b\܁'-xq3;לu|k+7o s9=гg6$f7%0v\Y\;1+T'ҮFL֞5\bdX1uƸAP-[u6[8q ׾Y#u1<\WE!"|NP̀U&u$>SZ#S퇝`.w=;b.`LekDfD%X+KVh0Tn%suY:=f-mWJVm,0; S n&۶%/N(t.rmlllRtƣ ڋNnwں;4(I3rDA,ɫ>!%[ɄRG};DߘDl*+W`CdV0[R|P`S$2) d Z]Rh⭯c( FR X0ZNWUg3l>t]d~9 ͑D\oF0Φm~@;¥ r M3iyTA|k1/aJjԮ pa4&2z ?9,|u5KД#X{SrlRLٱ#'q#&$$.+$QO5bLމpF*5~Fv[t[&G#*cGqR1LVCx+I?p`O[#//uhک-4h{%4c9qވ3"R&8EOw l|%mN7r)]|)z X9sJt[S]ut!:KL^LĚh52n$"z"3 ,7$MhIv_M8"48vؑ]mGUnـ譓f$5%V Vk혒J m9s2 G-u*ƾ ʢ#3IQQOmI 3dSEEr11vvݪ9-L68 r_)64ݥd "_|>8'tLqDGO~=}|Ne_$aZk'gG=@R/j+m#G_vvxŬg%I^0ih ܴb AFcp2ӂMrUhF}kv,6n"1zex?Be0:,ŷq+>޵o,wυZ]?ÿ3=P7781p`N`RQ6MX7fC8T5g?,5'FJQB]A*wg)nOd<:*_!1sj1\ >yNEv+%!ON>nUav߷~ڎZ ؎Jnfjp_T)‘ gX g*[L0&$ 藀R#Ϡ ڡ㪋lyZ`?IN e3ɋ cn7Q~T}\ɸnh\{}M:|IiqU;AoX3#8y]#*=8,p5䧤ȗXGzIWv ~Sۚ+m"%ܗYUZStBj߷=o}Z`jV)\^vxGG W# V1܍b(5bU ߧTn>6d23_RjS)-JTiP*VD3Axu3'%*GBKdGcb1RR9 nD&AUI !#zjMvdkǴ҄'ְ'Fc@].po$JvX1,q6 CIjW\5Ҽ>Z#)5;34X׭ra9Og AҐt(RfK"IrXڢu߿ z {\b%im~O~9ٝT2D6P >罥Behu,2s ilf]دnn?@0)htT-X$9O5  w C$0D[FTH6L"8hqh%7;/ךд_SlpelPs#E;=WޟG^ܜ1M\| n+?%(|qM*ǀW)prruKY}'K(k{ƕ7~iH`,s8MQI3|LdFccJ/M(gTFpO)Cփ cEѤ$R)pDz4gl TD);ᅍ}ᅦO 屍5pI쩺JsvwBM3}eSDD$y"4`BS|0QR.ld 8 Ys=,MtP+5Q'T ª M'#Mc~/)µ[ڶam["ؕ+wpF*U`j9HD"Z( *G֥.4TopKȐ Ԣ& ,=CIFP9a$[N#6>!Mri{-Gl9Ș+BOQhtxmzX!wȉc%YJHmNHmQ,e>P*P9 !$:;ME`>4H1rsěsO/'[gc|6i[2,VT,FI Db>e`Tj%O&hgS5%qo#dF|zc"Q?n׺CRDş7 ^,PYDYi֥01UjĠi|1Gp>VszT_||Ix쁒X-e8kvK :FhVs9dY7/|05u45B ;:(}ԖhIl7FΆuOiw̳bgYh]d1.XH-!60TI;f(ǔjbպ=;[ a P'!Z:m*"3P<sDP꒱Bk4w8;5/ݏՂ…{=#Tnݬ uڀ4A$2KЄo  EkG@(eyB];Y^hWsY.YrWcr .r1T%AT9@M3ۍFzTV I EOzZ\rCͩUӍil3*$ril.\Jy3 (`\&%)yg9>$L0wqcS\>Zr^vbnpP%Y">9g>Q>HyټcG2@- 3_e}eϕ.}.$py߶͎|ĆeN\l>Th`RjJ@UFҩ>Ue4.m=MAeIEcUR.K ?D5S,rIQJж+14;Sl6Y1 d+M(7e`n"@{Ђo[UNɨ'qre V_s]W߮bn?3XܜCa,nG<bmwO_{Moˁ[ͥQ FHuM"N?91RjGf@*ܗg0*?E˚ʽ&Rr})Uu1ҥQzp!MzaUjQV֢ˮ]w@W25/l[zvs|Dp'ni&n ? LT+ pP<>*oK7|wp*:5\l{-"WOQKR;Q朡`L^PL>d9Sy p18,YN~YXW]drQOtZ~Hq,I_osٸ7JuG㒷UI5)-'z;(N޼*ꓳ:2rXn٩aEuzJjQvX;Q T[K,b𳽊}17Dk8x+'V'ne#ŒZ:,@kVeMdx)>d8 soek-KƌdYŹYR4ш#{fdUU\fi.L+tYM-ceȦx25Pc0U\3&ΎizWB>r4=ϯ=A9+;!^]_-)GCSjA\%mMk(b(=RQ(˹jp&m9$k%QԱo+C$NU'רA";kp7qk<qM^~S,"6gLT[ay÷w*kuTQ@  :\eݑ57Gw#L'Օ W=/*LJ9&WCRl)śS凞DR1ZeT)lsyJH::Kw&Y(Z@&N$W5K>]GZ,N2JМ[bG WvYBMX4Ձ&YH#(Ȩ)ǣ)21*-Ovu9 IHX6T9`t,H>Po uh}HY~SDe C`9R-*@[Z"ۢ}Ld;zJ)֫c-X.@EoX,Pgt4a&XMX.KutlZ^WktN`[ZwOӁ4h'}:PԬ4]+&G€$:Cڹ 3M#D8>#%a#&]+Lr(mϰ E-rmcۓġB5R!R" KEvsnl' C.N)׍G#['7jFA^)*zH 1h FQOƂ"Uk2yn;WU5WJB!5Q5VJ>Șܶ--kmP}[G%*Qk)-:`[yJBm1 lVKըHugG=ǤfqոrhVt 'Xz|bTq^˘2NڲcɛXU |~- ;sϩVh[.4kw aƔrj)>ZJPlnZݨV<OzI˟F%DC2hD>W DX̂,;inK$v48G˫OYJ$cF6uWVNYVKLdXq"-ձ3EiT9!Z"*]1ScMA;}u=N^͚alsy)i2̛J̉Dڐ0n6y&^R˨Ngr`Cjs`i"@HAeqH]1c_sG<σo?s bUd]LXy[ xS&U)4OYAF&X^EqkPqLuPMuLZ' zv<" `r0YIăvvroϮ8;Lڿ :mqy,֔=`bB-آ#/%bW,m늌.x.SRm #fq Sk }-N;( 72vgEݸJ7,62 ȵD.3ްpqrrB,/gE!]|9;]rĆɌƢHA&'MBMt%Lڶ hll}@9{2)09,xNb@Hse}]ަ8;]k.池vѱ/jSgԦG{Fc,h벭IAWkjE) SV:љtv-J:s R! 3(^t6 ֔ -a(9Sa[T{C^7qܩoi6`<D5>BDTOD\oD7F84s‚ޱ/^#N!C&Z l니R-Uj_$nq#pÙTRU腜Q&zRJ8*: .Άr\gQ/.θG\qލ> j # AkR+Eм%爋ţ 5:C茇7Df]֏sSl/EmzD7i#YpxY],ۡM|p=vCZن̊d Hq7cbD`㦹oFt=G3M^\mr8Lgm \uu5AJ65v UzKfcs:fPIWofٔ(xe#D%7|KZ2]N=%SNyDR5[ElBs2drj9ɵmO6?p~qMvd8?;7'Jmv_h^g9jŃ(mҍk72{.}m{p|j&!}*.ʖe]oTyWYSv?Zgca&ii_~n.n>off5/r/qW,]q*~~}./~mJQ埰=NZJb#\P/E/9eDYsgs;uC5|ɂٽaɍɁFR|%1:ɹ4t&va,:y-?Hfb,6,{L~Y_ܸN7?Yu>dr|y_:YT F؛Hۙ130oh|E4,<,I^?8s9|RT%sQbyQ&3:c8ޞ0J[úJ: b',Bb]ʐPvMJ ,E *6UZa"6ovu@H8h+RMdk `E1+[1qbĤun){Ş;Wqv6,A/d<~\ߍjoW%6{ M_Z/ ׺DMcƻ D@Lr rUiȖknUkPz;+s2 ޳MFa^|ꪱmtEJ(cч[dm9P}K͜KA_Y]@@(g#%퓶XK6Q`0nJ tD lk7<>]݆I4u,?oxj#Jk1MoUb[A"$E/TJeD>s1|_FzFymcBrR.8E !7@)uE&S- *^y_w6G'xX+BHIE^\qOf񼤋S9L|qu^~Z?{WI ׎RFiGfdžR)afw6ݩa@]fʪ|Nd[$m¾𥺐C=^o7^?^ofcI56ޠS3zJ}w|bOէ~[/m&cZC(Df"G(*h*Ħ ͊DrJ.﷼"tim6=˚˩gU=!=|(f4($=&[k~KYҥ9&=~|/|Ǽn#se[6q&LS&*@fcbZ~j9jy2ȱ[Ȯr"&RHVXq(gjVL Q0CղEz-.amTܜ6y](L]st*4oɷهm/껞"Sz✭Z_kQO9sVW%q/Kq6P\ױX4ȁ_ R04Kj9'_O)CsJ95-[TS<8pɳ@ ^?zߓ]t٠)V`y"&&z+>.fK,C&]Sl:cL>I)1fe]r{ lNRZ*Z}%BP1̾V|l*Q$R[XՑX"iZ` .U)mBbLPLZc4E'DU,&yH|{D&[rXVKi=E~\lV}Wo#48q(5S>\ 4[ b<GRn*Xkaзc*Ȫt";GKTzI2ce9[NLAHٟHQl\a]Sf%+I5%eٍ"f "Eɦ^%nbBc5B}&S\ fԐMcӏ W1rnVɍBD~Ys\cwۧbJQk:I 8 $l6 JX#L(Hclb촖X~_X{Ӂ?m{&$ 0k2lϷ Mh{=l}= D<ѷC7-٢o4+dP<۠,5_ )>L>;Qm HcWMW! B-=ģocu&1`+&%cL+R](W0O. av:q5VBGޅC^ >(+`i|BPh/:۰ ԲrL3Uɐ5ʮ̑;,rus}05G/˜24ˍ8PF*0շ@_E)ӂqud@/V!b_y F&k(-\ alg>Y2QެpAL^fO(W͒@_8#>(y k |s%%9Fk_zNrt`\J?f5Q$(xVV guxg5jY':'g5J!?Rz8B6TV[Sd!P9_/u'G93p e=zHXBj7jݷB~w1/'l ]qNf (@e}UoH][R2**&@ISvt"Q۾Vۤ/z/7Ev >|KLdxy'|aoMErp(2y`x\|ߗ!F=N~]|iuʴ>G!כ"=2!aN6ydșUMMY.r=ԇp!2QZ- MrPs1XD]|6թE/6;LB1qTMj6} U1]X6X]ƅl"p4W.jP(&־zuVa*-^VkF$ 9}_Lvef󽐽7ӹW%m5x)ʭ媒#09g4-iMFU{wB(*K6(IkT*gl A##gB~<]uTZZO,ȫm#NɎ:+!Nw.8;jA$*z}bwɳUÍrQ^kewn{[dESjf19 M<9b,rK耒GA$c-KCD}I;FiBZXTIKl%WaUV uEppp!O_.>/wqw0,'JYz M6I7i$p' `o0nC-K_/ڊo=T^"r㫪[L+]zGNEBV%c)9l(θi޸wd#eHjؾg0~?Oߖ#Fne1)pqvYCÙa٘ 95Wǖᛛ[}lc"e|򆕜 KǓg'WmAĹS_˒QhKFM2lY ؔ(Y0aur#փ-<6߷T2.8# D込pg`x?Bs9"v8P'Gb- Trɨb(0@5ڌXӛIBp48C` RmYa1a< 'H{7_މjmL|ڢ^u:vMi}{njm,M7<1ML^s^5T'saCH EI+mM,3ӳO23#[C{E(xu)gEH A*ȹJVR4,-3#t=o-1p|xaxQoG&Lܢ`jjyK*q|H!TCT7â~/MЯ&ġ%R4Z>򹞽qTlTT*bY/3[CS"L IY9xH"c$r1 6{ZAo2~p++:f !TE sC:<]K p=}RAU!U~̒J=>љVez1DI/+^T  *6; sHфɗʗU[i-4X3!'+:6i=e(PW1<LJto!k6dka`H%Wwϩc(^vÛAYZ9B g )/(Ϩq6CzixT*ƶ䁄nV.><ptT_(glKn"'Jќdsk,Sp|bCΫk?[gm2ܲ]yc*bTJi MK#vZer#9SSfhy(P *5&x8RȰ,40XX+H!Mty\Btpͷ6ie+ Mf#5dG {;gOE.zI16/Մ#ȜKAa,T*#Q;yzbWqK,O1%tDQ]褉D(F4J@7W2j` J 8`1wa%{:aFhNPƦ~6gI6%,8çgeģg۴c?d{Xm[`ʸLBL'I^ʫUԙ~VI74s[EO);} 6%wT9;37~ WїK>*7;gC g7Xt0uQ k)b^3]HfG|xA]]$=B9|?e'OU>]}f:~->+E@՟fe0[i/UV}ZuĊP%%(x%0vJ#橥j* )G1(fJdJD!e!x0k5f,`ZFL&Z ihXn Κ׬ųygQԙj}j^y@Vz%tIz0?1Jݣ"Ht>$5,{ |^ͳMQ5ϕ\=FO9s~l ˳{0"Ďds.~LͺuՕ/:<|Qo]'mnGlgSKBsL{2%2Fh\7WZ/֥]5uec\)s-u"H(ew/ԟRح=Kή>a y}xƥ'M[vw8##GJI4rr{̨bCQQr A2U1kQ6>%lK718k o2gOHz.z~<eszjïX!ˀ B6ik:?y4ZHO R ̀EFE+ C¤+##$5" |nÃorBorDJNc:!s2JaQiƝVc +"XA%ߢ>pbBaʓÌźVW΍v2HDU:d<*6ˈw/e\[Za5J~6&jIJ)DO1HaW Pac,iʄn!5q}nl \ ;?!ک?XRp"4I>6L"r&Gb`XqR |l0_埣&aO2]xe2ٿ ; dB#L9FRP;6q/ qW  N[lB(R0\_Cz., I8흟ln Kɧ%6T={VVW0m(0טحU ^Yic6m0:RacЦtVOfF0V~epv a\~[g};g(6fR\}}iChɛDy71𛮖, K_3?Y>Ba0,8 z0bɇdzwzUG֍\em)>}8dzs>\f d?8'UZz/*-2NQ-_na.ERdž(l?qHFr3𙯿A GWm8sUٵD銦 M ʏ3׻Epwo~8O;sL뿿> 2 .eZ^ A ᷝ{/뿚^ _M[Սuޫl{+_Pb;A`]m:$ [dRGZ&6 9w~WfKT>1(PQi"FV78ƾKdAF=iIM_j•:g#l$"ֈ+ݐbD=@ q^In锺綎LK!YHj q:% F\cA{̂@/f0|TC!T3P9eL8ʚ WշƶQ߰t ;N7VoEY[G_ ܿ 3*g  d;O]⭱犷ƞ;{xkkÖbbfBw/r2kBByb-؅m&.phxXt8u)9߹`LY'?&g)5`9YĔ6 r-,KX:y\EelfV۷v'jMaZ5zdx) F\7PqB$&4MW%4MB;S@C$ti^`h-$!Jpqk"i%*d]$HW^[DW 3Dk\ R|J;ztT5t2Jh8tJ(ꛡ+ЋcTu99=chÉj .%f(<,Hz롗=s1'=7.b@yM9ǃf5jC`RGSJmΌ6>)Ȥ %rӴJb¹`1X>].3V9On>Ɍ]6lɫYGH(򹖜N)bxЄ)YDDBLjH;Fh.y[S'1"Jl ]]%:]%tt2$l"&)Tk3i R˳Yާs. ^D#nϵAsϙ87ʹS*EgO1h21wW-T!V?yU̻F d gG;|Ju3 fZ"7LAExH1b)E_ ;ûtLjZRJt9턒eBZc]SտGOWG2O}荙xsO^stEZ~5ۣM'KU< -50`Q :%c1F1sɭ`Z$Ť˃o,(+tRN>y G%-9:s"9cn\w""{*'LFr'pQ`.u5r$WPD[HSĠcB)&x9XGEdQ0A[R.k%)$t#׿eP&u$d %1I,l eFl@wW]*H6h6S(;?ZXuuG9in۲z3[zV@ 6Gc~W3S<}^4NAbaߒbŬ9}s)k^bnטgc,~4 ,( @EB'$e(t=L'/QH"8R&:hHBYZ&nL7RcY|iCf%B@+%zUa&g=CZ/? v9Nox.izכ65GO# ?SdQ5c9+錳6Qxy-FVr0VB(a(6\S  %z-!m &G|J4*F!dp5tk^rXӀڻկ_zt(eFaoysrnMREc*ċ115FyHðev㶻ݥFdwFvۿaY`VJ#ByJ)dKΒOA,.I[GƒmQ)6sc˅dAX@ǙH&КHq"o>xmn7L*peɌ s:4;-O.K]btr L [ϑl6fTuYi,+MuEؔ8M)*i|:RX,҄. @ƺl:'M GSXu*'d1{(92Iw%fdRD^GK,N#d$A!3dvjs82x^\ozV٠'HK(F `$b kgU/]U:ؘt MƇhW eU # h^d/*Fo$9Q.L:6"[+˱2?S/T;pmUwg|#w o>V]f>Kl{&L M z#GMwF7-IGBN.$H"=j]]|/6*]wɜ"hvX:wNk(Xm)Y_*!AJ'Lƒ&5})皎p\Fk4 4b~ѨZXY3tY+J`ȒM*R~ 8>B@3*K&48:디VDM;QfT+o&;i~~6rnC~%Ԋ.*EJk""F-dM[݁p4UY}!l ٗēGb_C;}3 $$jg,ӠLHyjJX^}zj˘3Ҥ{e L!{:>/ EƢ|D\3o#ðD3%PJ+C $y 'j""8Xv_ yUo.l#nMmoHJO$j. a 6y`NwTN؇wSA"5uFHMWS_HZk _"n\n|VQq pJ6uP:eRFɘ3r/XGvgx\^!sŇas*zTP٢LBld0@5A6=BkؕROzR'LTS <,Obp"T &ٺ8o6b>=(RKI--K5rƁ͢?H" "jUl*R)BFBjj[-YZ25B. ԿD!XI&3{ 6d>j:'xѮ}\˲z1bJj{}î,8nŝ[pOXv=Lqꑡ ^}EtY5uI DA+0c W61_D_e")%kH $C0X Z*+skdl6r4*Ͱd 3\@8by*o]YVA.HBS9A`j%褴S$ h۶ބc^D)TJR ʺlu m;Ir!o]ot#vZ9n&vhaBxShͦTZUC NYbV% s[<$c]ɳJ4[YXEd+ƒdbB2(22ͦs{=YΰT` "I߄"+[erҨ%|hKw$el1:\~EY[D bzTq&alPbkb+3ռNE:˔[6l6x]U'UU1:Iɹq&\p`d6tA:j>L|dwDt&`;kh[Iǹx(<_J~Y˧~bS x6F2я\nGyMm i,BL IS\S4hȓC`\|~6~<8or|z9mos^Pc.7de2Z J7NѤ wpQ!RorCkp=ӆQ>z֊`:XsV󚀥'c..~r}kЫ߼/D{o=+KR)}d-z8t[ gE:ex?}^|uLJf<46<>#jluߔ?yQ=$6w]Կ#c9١A}ΝQ4*WEkѥv/`;@uik~OϠyz,l2iNWH bɉ2 AHYCZJ4Bgfvr V,]l㬯ҾMf ަeѽ[FBr61IVB.PZِd둑`X-"TY_ ~`pT Θ79w%[!0Ѻp,"+grQ3h j]!RZg2@@4JZ%]YI S'b`h60OƄaD8Wk݊|>!v7-wS}xF|pw3vx!ǚ]د6NDDh3&&ki$KH8OY,8A虐gC哵2e,ȠcERbnJ"T":"zE3J#1x4AJcrlqok$萘[dDŁd[^6q .h-_d+J7Q֓Z qdؤfa}1F|W6_`'hLۅX5:<8dH`\.SlOj)2g{Ix?>&u#?Oe+}qصQ2R/Z>1?=]ʁ.]. ݔ^*/cQO6o^|]Xԧg Z.>]\ןӇtѯR&jŇnXpwt&zxus6SY.X-U 1X`YoV,h/ºmk=fY nllpۏa;d@]zZiRo+eM^\nZV7CcؗaNC }im,5@ᔁ=fil}8?MZc}<Бڤ;BpȄm>`e ‹~8 7VBtI6NIy 1Ckd:`V{_y [[Rq`F[QsHOۼޝv9ے+Oi{~w:sm[#A:6xn{j#&~sښ7ˁvv~GoD Sre>FOMI'1䨘Xj9P"_ D ]B5}D*iQT4*'DjL!4qsM}{^HkV4*kA&\b.lFB\lNdރvZЫUfqnV7myyf(#%a%?Ybyt 7\ʲ7ss1:uޫNGjlIKRRr7 Kk;>pH}l6_k m$Rn5,煠ID(yfPgi%{6,5f37T^r`F:Z,fp+zA^X0 c")˘׊hPudx3I< zq41j[i\ʄh*d .0 Su{SAkDPLPԃ!"pǙ3\8UDs(MXnιo3THY"!(OHceښ8[M>Tm?Ó=oꤺΌ'솪'U3j.Zњ,]ui3yc|o}4 -Ɛlְr4D/lj ۈּPr|k>s2[A?mKsRB-7LiEW s޶A4sWIW[n6!ۜHںm>]_y-k܆>RbU`P9UIJ-co;v'Ԏ$nϮϫ5^djՇKO02Y[+βvL1I'!?2۟kk*#!9&K"Kh&k P\@TY酐e!v }n.>3)~^cbgff'.9ihh]u{aB'Qb>ǝuSF$ Y Loߟ3Q#) ۦs˘L;p0J('֊sƨ/cTgAh䍣АeO38N+~":[&? WWs(U?u4)icF>w0LH= OflANW:3۟ n/ BVj{[.V5V7f'G\!l(m{7Ӂ^t:WwN;U}4Tu|k窯FR'Ǵ3hn:a~4N]-:^}y#cnxM"9%^̇:a':~R]HE*7]߽/|~#vU>cim.\eF'q@]q-şD5._C] $, zؿ>ZX(_2Lbv7/d߀2@JQ&AP)fq9*,{*,!JR)jCBq)9,Sφǫ,%/̯Q]UXQuWCQWYZ]]ᄲ^2ߔ?Q}[*+]eiwuԺoF]e0A/T:ˇ\M~O:s셉'~/:p0nȊe'P}̴1F h׻[)5xzwFl>.q'HspY$dЏ+=8MccfT  i\}>ھc?Xx(ins)aɁY`:%6qε1de|˧XZE*R%<1TI`NG8}p&EWE^ARJ" ePσi ;uz( ]]e)(+i, E]eiwuR1E]Fu%#ޠfuL$si<-a-Q~8{; ޘ ԧSϻAL =\KcV:yYieJrD{ߜ =1% ${Y2`N&*O~|fVHmzpYpL~{Ťv$Å9k̃H- -aX'۞Խ@ۆFfsPlͻ*jh]zp8n~8QX]>رŏ=[LDq(]WL[J $mFO ,mp$'Ϭ1c&>&{WZǟ=c8Zvx =%f$ '0=ܭ^>(U m.G[-ݧV'V^/8̙@򬍤7b8U:Vߝ Sb(+FQ>w|E;K0V{ŔFgB9D=OSo3]L /5q9+rƾZ.F# #]7'!狖[_6Wy2_е8]?y5gX,)PB/ڰ;W%Z$TFKQyfe4LXY,̱فdoQz0k` 49ܱG8f {8Rg-&ji.j' @ @4>Yn=PX!}BSo󷳸aF/ No~\~gw?:YkZW?+"2?*x?"xyUgiɝB ׽v&j^mt)gVZ A*0T.?q%U5wJhH9wFUƆTeHgB肣 uDªUPVAYQiW@-7ۂ^iI"X932ڀT2vUT )B*m*nBz&eIyV|sɾ:) Lo+$JftH,!P 5t\̭w]U}fseQ ALJA>B QC}ťP\;QRB} dy y6 ~ku.[Y ,TPWhR:Bb-Pb-kİDkeytFGNhÂC8hiQPV&Fb$jg>aq4 Uk -%8xI)&ႩDQ -H!QDLrkl1HOF@GVݽWRH@q47 H,s@:;/CVWfGvvi%ė];.9yz8->,Jx?v͡b 1ɛ1%% YFeB(a(+mjɃbX<z +ޤX KQ `&b .7Y߁j װP`O +@+”g\jhs=K *r7uSgk[ÁG 6wp+ÁlfZnC XKi&@Lz@9@}C}iKAjx٠MXg{OekOpҵU@ogvj %DH]QX\Bx),3Fh0Յq3qlf-V BwF`nܽPK P(Z%`)IdΛ >'c𢔨`,w̬fV뎿; B0}*v/F cmn? *'vJlV*^͛lKWl})I2N GcQ"xΚsuv9ߓ2\O/)*}6oyd%L-tV"ellE0*dkhi_81cPH2;M(c1*u1*(K6(C6pZͨVZ\~'} jSӨy$AkzO6) 0D D,; Wݞp4yY>6zxjMtQ6W۾7仛v z5}M3ciP;k |?5EUw_qVm9RbZ5s$,(`ONN#",ݬo~s1f*rƷ~L ð4;]Y&釢IT% حтBN R=gRxlS1lK RёJeҀdRkϮ8G/ggZZf5h 7ӓC۬EVs7,'{Y']C`:_Zw/oH*QD*y6IF NԑV`|:؜jr^SdDg)gQ\&%LeQP#*+Skdl&-qfXL3 ݀/-*3>\~5<"vyVQ7y8\_OﯯC RHgV d5@l@V&BӸK:{!KbJVTh:[%$/#B"YzW+q#vNy.L;NEmݫ2j:d`0U jU)6TY蘄)(F*-ARtlR!33^tR2""kC(3,;ņ y9S_,ͮrx6J?NBD-qހ"FB C:kd!gRL}6F:HRȟED@Z X j]wZRřcb >z.%mT t1nLÈx𣎌ʸ)LKNEh0​{m&uXZ蠙0 ނdBzK뀋/gSJ;NCixSy(7yTmٿ4{r7IO(텽m*j%:hd*gXϥߍFor}O]G|638n }VKxp꿈#h;=_<)1nOz7/*-~?~}OxnTz˳]ʁzm.;bbϋsl3u;~\wn{y[ym|7ܔ !e1ILa"JYE2BMX+ By}@FPu DgmtEIYB ES`[2-$2d)$Xgk4lJ .`tII SOy604V1sBFS.VbGiL>}8.G_Z{Dxu=5s=Ou;p^:FvND 5Dh#3&&6ze1xŇ<^{C")̠'H1Jle,R34T12{'b&kŀy8@q^r]4GYQR ٗ]zSQ:#(BB!T?1Ԣf-C%C*Rڳ ! ^ xuYNiX9;0y2 $o4DV{P!/zQ?n 0~JT0I:˞wɓEC L.za%~2*h /7]B2CdއFu?'RVJ]v25WE ~rG XR,Jl.i1UGܣWG?/!}$Y]貸(lz3Z|⣛ny'fG,^gkˋՏWnl~Q~[fL1k**7@CgRάh^_fϿNGbnI:tl,l/B=mwXu kc/Z@]*^%IWb=RFmHO\YUpw|9Bu_ޝD- j_̪{pOΣ9ϲȤ;֓D[xeYH,bwu!j|o׷Zz8ܯ _,ޗVKMGwNǞbxf\=֋9OTNja(M~e3ܙ/oS/|QV(ىiҁGUߧezuϔVRՏkJlDNՒ#tYu61ovf#&pwuvNuJ37}5foߎQc Rwq#wZ ,!|ߊ`rnl˅Wol ^.8\<*ʨ[}1,s݋F\{ZܝOO`?|4n)||1+yݤ]%A-C)vK{Nx*L3/W9WKw9ͿA=7fξa?vH'5ۿ:6X1/wk]}ڶ/_XUu:z踲Yj _󥚸e[h][um|+Z[h阮WwnU_ 9GgCdϓ I!y䊁e.dJ*jQPa&oCyzsNl E>ȧڸ0 mLA 2.=zLqr`Ő&\<6ӍS#3Ֆ=VᎌOCiwa?o# FFU\4ocv6bd)C#E +oIbfٻ8+W#^acfD$CRA=)6g8lZX1p֭8Nhwiy?ϻeW :'?5N*iiuKJ/CTz\tywrw(.Fp ι08XM:vW꛵>EL>2PK64;śL#:wNMƑ^=O.Z.7ʵ9kdxcʓ~-7"'o˓ Yr6_O5aZxW̽8wÄ@yEC{oQ~+Ż?w5|JMY5WY`n?{[ƭxʆʁUO _{=oo,1=\#zmi¸OR ATR_zlaTrIG|0g}kjKb~yY+g2IAغh]M^** jT/ߝU/}Ur>_ȇ*<}(*I'pwr_rKw>v[+L|A0Xy1lIXij%T/NINo~x/V0:㵼 SBjIZep{M YN<֒2XߖpcKc,mSqXM HQxix2ǃ|?A=@ۿ9`@(s?]jxaց뉁aoy=6C\}*f 3T>}@zL?giJ}$^I!ve9`zEfcE!9AB[f3-9Wgp魅-jf k4_j2EG+fJ ݕ"JiS0]Sdޝ"7U^/(/nӠ/i)Kz94\^R([iJ%>/g)48@H!7h^ͫ5qBy'?.F?lGxtq1i!̡чb׋<~,\Jаz?l-Ud yU_QfJ?JgWҡW齧Z%+ z^N Ͻ\Yk=r+5+g/)i_\ēN?s`v|v\q9Q .>d.!׋ GATv_hJ<2}v["{j}fvf'_\MN@Jmƛ\?+p˝w׋_i '|\כܺ|A_zñ kpG?#fꮃLVyU pq /s!ᖉ<|_x}sw̮2ɼ2tI!6k$WqX)Ԩ7VD#JҦ@VTeVfkQ,&{ꠟ:ɔ?9@ e*nY;Ou $KuZ5<  J9ZـH;69xZoVM8 d3u&l |뗃u+k^LKQDFI1c(QBu"3C:I "DT 3"vv3lL2gZ21kTд-YwF<.Ti>j!0pxj>%NyYxU2z(9@mi>j\U18Q|ŝEOpE A-)JI VWed䓢Z0´0G^ />$ܗTd2jnufǣf@܆`m]_g,XNU~t}%*y*ΈJʡ l+dəj&@D*ofj'y's8-fNXk#p`Whcѣ.)1H/f9 hT`%P\Ha(Ft3 S@icJYI9h @;`X#F@bsI.c5HwѩzPBJ@R9$tiPOW,X7oc2llP '1+ MJ*UJ\ \DޠϺ]FLPiQ0RSdLY$z΢i,+~(]ڈr!c̩ X;uF&AH@pYPAR䳗;d$0r$jd-ʵNiZ-]4%S#x3*"aw*>@6`(…YaFY&]@-L0E zQ{!1?M(h ANr|dNrO=j8%oCI3I V̏׈ LCicYc6FJRD$bp(d3>FPCt.]XzǕ-0 ]Z2ˏ]KvE#B@xPR I0@5ȭzFO^tzzqkmRǐLY%% *]w6~C h^}.@-Pg;ن|`5'iY\x?! -a/Fm kzjKXBW[-%T+++++++++++++++++++++++*Λ(/E ֚gb+JרJJJJJJJJJJJJJJJJJJJJJJJW2t'_Ν@0(`m^ Z-@R У+++++++++++++++++++++++^%)@C}9J kQv{@v%+T)f++++++ٻ6$aGC?x 5 Z)D R}goI#'@d_U@CuH !:$P@CuH !:$P@CuH !:$P@CuH !:$Pv@kdzI{Siu^0o/NA $_p DR"Rv7>7p)QY\K+4$^R5 Sq-+KW@-EG_q-QI:qMdt0L$XO_\zkFEbKxX*A'^:3Svhxg\VS o ÙgK\),>)EOߛ߄>,ir )L]_)lCVZ"Hw$8 $&Jxo H)i_ur:"AC1Ex->.)h^qݿր^O3im}x?o]#3hmlc$`Smn7E}|իWMݜY~SU ŋf7G%=7{IYU[Tk°pN5L?*ie,:KK-.C߄2܃0c`)2 Ō93˭"IF jmZnKp&x/9tw#7<ށ qηCA3NԈ,>˳f޿ 8w -+`nՍ7uY> 6?lIΥ/7zefW!H=O>)z߆E.e|I+|ջ3·OWI0uqV,qVgEAc-!W˫ozVMz -kM- ȉ7w:)McYgKISG>f?Mw>k~M. %:>V?O@fz-=?zWnˏV5ĮOwζ\=a噶Ix1D./=@@%'{_[!m y70ӯgߪ@ˇ'IEeݑ&[[k Fl QĎ{J e# ,.kh)iX,&!:ҠͰm6|6ڄLΊ%+hdK?;lA_1К6kbDb1vδAx;nA^OAK?xwEl<76}/ħEp0.Q&Z5g \euDkCH)z|R85T+HJ' Σ`Q``1G P;кDH\q': AH  H% G̍)MI J3:Ҫ`h#[5rv z{L@n &j~W&voJ5gۻI"xlT~0FW.xEI0{ZǹjN7ƠewICNC&A \B?e`Mg'p4CHhN]^4yǼy:(cg"!Ju7PH(qjRhIET Sסit&2%>D3//I1%KS5u0/c)~)wer_5:yIt*RV=Z(a̠w_%o{G MJ&o"G1rѸ)ݼKQ 7ukLBuQ>oAS6]S|g[(NfW?UQ{Pj L:XnZ LB,h ӧg?/UPpck6v2RXxctYE5~"_=;"huN8[mx*{LR&KuG)M]=2 ]hwh;\nx32(֬#~fy- T3Πo=DeɰO[e_;:YDD~oPMrӓ3/O[_'Z L2:|ݚRԫ+zW_)N7eJq9K9N^`% ƊUrj Wp b0y[SݺW5OnFt^=|f2?xCu1y/z "`ɸ,?5Ca$XHoi< ,20UQa>4]b>f?69<9:dӨMs)>H0#UEOQY`lR?'7\:h.J5foorCp)b46'W\OGQb+=X =Ww.]+02Ug5~^=l7叟/uoo/¿{4& ֆ﷿?4Em M3t|quSnqqݍ/Ea@KPPqƚ]/}5_4]}ɠ?+à z#0빓<.<1qZl(Z#`uC 6: v 1N%㼒uOJ¬꟏C"ju؈)1t0:0s"0HMΈ Y ^U$Pw %Qa"Zb"(kQ:zA}u2sA⭞'JQXP "g4`)$T@9,%& PdCS:gOtα9l)&am +'C p^*EkvBdhx\Q:>*ET7|͘y㛧/gfqYs,=RFA{(.c|;޾;vlJTjzӥnɒX0}cξy#yxRdf nG~V F<4f0t0cg|V-kEKV B Q+-0G'M$`р8ZX?Q4P9:=V)n)sV!Uzro h)aeE5r)jVcRUg|V< ߀T!΅t-〵?|x)Ď9_V`*uQgWL'c5U+It('Agr'8NMVT٩GFURJWied;Rb5JYg ,4f0oUM!D!eUŠy0k5f,`ZFLg5Z ihYMm=c* GÆh*Ӽ%n5|ȴѧ 6Tn2 ˪O~0j)y\jg.7E"j·Zp'RX6Xnjq3mΨ2CwY[ΝxDy(CLJ9%&eLrJ18o'GZ& -IhEܧEg ?5zY,g"yz%+>{q[P?Iٻ6r$W>N@#;s؈ic@lR#ʖ7AQ4%EA)UH&2^Y/+0" HX4  s]ƐESYjȓ$ 5%C좭ZUA* ]FQ{K5 ,jؓA*hr:.13*L?nyy8;Jȧ7t_vv߹ MvKa{ + }MXի6wYer1cw%mՉ^ _f9е^k%̸w[s5l! HrKu+8/ul}~\g$%Fxgr`Ķg E9"c|=" M\(J%$R Kv Յ{Ǹ9|f'i-[7XѴR?M˩rDm q2[_Ar`tF!ZkQ$Z*%ʨB \*FE2㬛9;5 Okht@N\GgM'@n4 zy ,* lud&c ϯW@NzI2D,/qVK 3ds0h1+0ZHQ|11OO)+Id T}.ZWpV jU6@&GSPGӳO)&ީk&ĚLq2R@t{j+蘳҄½CGA.z+"b[ݺ&ɍ^Jus}v-]2$oevx +p994C]C`-SMHȥǥ&ck Cnot"EXsnWS)kiW'{I$ZuAXlR&c0g3ךclΧX)ݨu>A۽6zY/~c!Oy:Plt 5ɡ!z`Dɝw燞T>2FPHY{ K6-,{ 䔚FR껟R hrV8 B5xջS g58&s19fo+PeSRdvnD~vU)BH=/? '1{d 5Ǝ4˯ ZIm%3g$) 4lp!hO1ڑ$=$eϠVy7F>iryzaiVWTflc]46.K d[;h6m[' ߵ~2g<%2iˮCe6x)edrr8'_ݛgg uwI;8H}o2JlmU\&s_|98mU}PkW7.1ޠSN#z!J% P*˓r1|a=4ɡcL0(r"F#zN->rҁeՒ- P YE|\F۪X-V(gs-Q5L` d:),&d3b-Dm]DKiJ`ϊyU1`*C)\{ng7svXfs@tG: OyZ uI bjyQR%]BUcRh+ZJː2+Pk߲T"IT`䵅Kr$A⵱!dYFܻ,9;ȟ/wVS&O,d%=n48&R.zm/gh.>yZ.+V2_\_vSr`PP@:>k,T %"Mzȡ=]𕹦B{>@X%JIs,j ;p&Xչ72v3gvݸJ7,agµEml3eSzZ|]U!N#v{%Ɍ Hi&' &PA:c l+OVA,ВZ 6ar^* T5RAr5[2aNoݼc_ԦΨM#j ^9`]%7)(jM)AV)$53/V|)HHI,PlᢲEAI F1q$w3a<\&)c/D?+7"∈tqPrBFg!At$MC YgMС*'ED TWT*{ԵL%tŵ5!$˽0{-fÈxȭNa\KE茋0∋L6KDWT-Q0p!>QK{ ;PTև-ɟ#.>.DR/ug<3@Za Es9G#z,RyUF,zܡn~|G G$S3^'05k]8tf%:LoQ)Hi Z1GW\:\Wo8"jw4p%:\5+7@vA^!ún>gcrsdJɯ2$ege&/թ,WSm'ڴ`I!A+7F4®/ӏ3}Y~"08+:,V jw:wnp}?]N}3N/EgM?'PW K;&Q geR}1kUm{k;kFK t'~70X+@F,s2Y|<=;VOR__2C^uo~~ygkPYѐljvt%\A{D,hffw2ۊ¡϶J ζyCSJ @GW\s4pլuYWo#+1X:s>j:pլ$7+t `hઙ̱U~Uč]\YQ]U3]ܩYFvino^n):3MMz@..7ccgay6l B\>.|:M sygikgkA=aLllݝ:-b~UW_cPb٧R ?>O[>2?'/԰u޿kouӸV7nc9d8Yx' n܏m6ل{.wjc}>C:F8BfVfF$b"1]+Xg QhO=Dvt3ϣL~տ?5޼v|qPz{.f_fI{tdQO>9BTݩޖbyy=FI# f; {by+Q&32cOZ`?w+>wgi+rdנ;֕uB`!eJԪUU0Q5hIy ;&HV,@H\iO鈺` V.cUY9;39O1g/_40_G:>iꄅyѳ\&Wb*fU@QW#]DPJ䊵muX]QB̈ς͗Wn< / W[_ܜ|⊱lt+cч[dm9Pq}[N7Jplw |5 L6͜a8!|fK 9:[y!L>x8Tu[vMa jyxfk+Teq9_$9Q)ux[1\ %/>H>Bψ< yLH锹SF l٨LZ5XT 4ɭ-/G]H) -8ObIK Sʤ1E1Mj%fw=C#C(R;;M('#; xu_!Eʼx0 O]Yo#G+ļPRG~ifvn{uCȳmyZ7H)IIC*yEFDF| xLkAi5Xm@_JN#٩O`OImVO o+ϝˍ|VMCd6Nd| +׿A;L*REz}(Rep5כJg_s |SLv625_C4j)C{7T^_ZI]wm&k> KRVRrLf j^%VO^UPBH'aZoQ,\4)ԡ\S"`Jܱ]Un1H*Z%yK :iڪ=SZDq]NgL{ҙx=e?7pa}$0 %߾y3ZbU]Ά/$mhs),G6Vc1~dTba4-;O]z@>sjn[=frBWgd\iZT8?u׿%Hl؟N}L@Ȼ>0 tBQ6J ';\-shuT;gj#ע2(ۼulTDU5b=0[+ue^;y, 5?îmjGn~rOU<۩n,1p0#h @xZx k_'//,+`6_f\^_V@Ls<+SHt('Ar@AJAفzt;-atH^#r+iUXl5R듚\b2l/-EU:~~ԭ]z66j^1QrzF[꛿^=<+Q7tO#WNz宥9W=7{뵟ߚnɾ'tTsdnjn4ko;p:k{D:{JD}_ @{M6Mҥ`_`oa`N`O8\=: ^r[ìg8M兯W4umA;b8MFSN% ] rߢ :xgc4?iotW 8IXfp*6$e=E~ zNty5e2ٿ& {]`@2N!\ܢRP0IE~7 E-6`M )L.D/0 }wUځK0X|X?F'?OwkJFRJNטt?t*Mst^`‹ß38+)C^>ե{UdfKsݯ՛_ogp¸\Au5T/Ƕa(6lZ^C6PKK֖RϦnښ!Ls |8*y`}oh8df?99:d[m})hu.3||~]z6{"~/poTvl%+[pZJ?9vO\ṟ㐕6d_(Jx{~h|S'=WWVTaVUuoaݏ~{۟_a_YWiyJ~6 >ŸA?o\MC{b4Ӯv^oIl(!&dJ ް M/ `bdHYT`#Zϝ䰩G(5 F7$hh2;1+ɭ1RwlȜVzNQs3!8-A F\c 8 QEhhuPy. 8ۚx"0ey}=*kVN' 2`ou+F8D 3_  U0 K)FD_.9`X``h`kÖbbfB/r2$5Bp Q^<igl5Inmjn~~-Md z 7k_ aR&j8FWGiHTK#&VnK{|R3Dԁ&z{ylf.m֩@S%|:Q1Qٸ G@P S՘amo2$n9A)y:fw$6`ô5Ɂ"O`RGTPJm r,R4ƺda߈mBZ|3u淝ˮƭYܟY㚇mg; N!e~aGRVHqK~8 }!%Xrk.sh.L432)甘9t] 7(xk=> B+볺XS1&xydZGEdX,w9lVJHMܰ b X@\\rJ>[ܫ'Er;7Cх! 2~mUTXiD7]SlkϽ 7aN|O<>3 -L 1`l,BʄN|c0*$04X͑EZF%h$L/W^ ѡ s m FcD2YL^jʈhA #(H8`3&"YFΞck\q;}FO8/o5]vvwUA5ADdTd^%:cEH, +-).+a9$)d-yA)ıH1WOpn?9%yѠ ٻȎWyJbH6 @v_,m,Mt=?VbyKOht$yͮR,rI7p M*C/$撄5i1]` #{hZ+]u,nqAV,).k3'W{ ݼ7WEUD,,17;n@g9@Ct!44lo =>AVC.3u[+f3 jRQ'ˊ%aj hi7 T Q ݳ+D"6+Rm+qwt>__4^uv7x[Wc&oN'&~y`t}k*qB?ZWmuNO`֟(5ѷ&vhmnKFh7:3={*\G/t-iqҗq#~G㓙ZA Sω[Mr,Ή9c9ɞ(P>MKh~oV_-gR`?oL-Q@KCHNJb?c5x c$[$ZeElȽkʩ֌}1P jУ(M('Gnt l+FvSw:58k8Nxn4#R)GFǝ[#N)T8M-qկKmߥbN,xL^SL5ȸL \3Ğz(BA+ 9!K ނ:2pAڏRZTcr@9oF,7PBZ]\rzIkZFe] !it*)EDDLP`uJI;=@ڸMNߘ]'7æ-f=qoc!Da>'Z2+f^K14^(^Ja?w!Έlgmӱ6d)MEv_|G$2o*=֚B蠥BuJN>yi]̵|l)uM+gl*G>.iȮH3#$0EÛ^2m^2]VϷ$eH˂䉊M>q&p"y6B"e\IK>.tH:. e5~{ߤ\9d_P8[P:MM^\7NTjik,޹6p);ͲK-R7AAIg-pD/1+J0:u(? Sko7E83Hk5#\,&^w uTbNs|6qq'e/ i?#-Q5-P$Hō#ىfUt]N1|=R Ԇ;I!Kis#jS_4!fc5q\fllE{ibQt;Y ܒZN.m6=1v+#s㏛̷䴌Z?پS[=yֺFqtr7 ?߾zEA2 ꔊ^Sr?wQrH*ߴ򨖙SxT0¼%rj !֗q6q22ΦKB^QJNφg|K9V|<)ao^Lzpp glIЂLjX8_) b 4kkyI0Zջ~fHT愕T9QJuW283v>:B̥lڱ+k̬-+kKFP! ΢U:Ŗa 9AIh{L%A"d -.qMm$0q=X]]!38&s>q18~Ĉι9VF\btCXRJ4ixakdˈܸmE3beji4Rk@A \m8837lx2]kIٴdW^yV^\yJY*HSsECE}Qo' 5QGd:깴cW>3ߍ®n*[]6\1s)8rZ4ڂ޳~<;?5m^N_6^gO{a-{w~;:lO=,ٝgT.cfBx6&.Iz.&taS[n=ٍU!GPi517LՆឋbx޿8׿|Kv7q?Ӏ"Ŀߛzds}}h~||ŕ8|YN?7>^/&3۸8`{W{'?`]T g/=mx_>6!:WSd v/zvX {[=U{&f#Ѯ/*&WU.ED} \[Ve. ҡCrrPE ԓ:/=E[LR˄}\?.^Ʋbn5:vd.]g-ҩvT$I^kJL 't!jep}Yo읧Y~7+zebT?CD9ǺW9X3X!]eSƓAQ'd+CSH|o8/ ڷ<~~3^mF {k[{FƞQW]K' '% -KE<2$7?՞͌wi5/bL+a} cpM{$́QM94`ɋь2sll38k̖q1 г?]ǯuߴB{K\ؼ/]/w gjcgU*%xdAjb\չ`Vi-ً@왳JjDTnZ w.{S >Ng~gZ_=ls1y9qZW>|fD)<$s$ q%<$6!eLI+CEsjmd9#Xk}%.n\7u]"(XTFH2 ֘]b˙crV=%BNFG5]҈]|3c~8 NQzptbMZL]FJe)G"1l0dYY9.6,͕zWn518[fy肬HYRLQ](9֖ǞB7Ub* D'={gMh݃s6ïi=0]RmeRc 48`~4c +![u왺jߪX4K@˒rg\mc5+Ɩ)GGW@k;8s_hqܰ]j[1 2l./kYy[dzz4uqci$RT℞yB)R%}&D'<;Z-"L%eOE Wđ qKd?1[h|2!B+..a9q In9|C1g,G41.Ŵk,ٗno]-gN`?oL-srQ@KCHNTq[DU'"蝸r(+rgC啇go]SNfsjP-x@iB9i"IBnW~c(e {Fh>E<c$o(ý'Z&,Uz軎Ϋjr~g 4SٱMfۼD2ej+`!k_Dڷ~ucAhyAMvpc^Eޫ{mv^M\V {lgcx'q*v HD^ZKp7]\jٿ[MBatY4hֿuE$s2b_٥=j;3hx,F UQOzw^hA Wc?FlJ~ۖcW‹Z 둈`B8l \,?%nP T 3#D.c~qF^aV/o oSDiđeP;S\/ɉ6Dĩoi 8Oђ\"*0GD%%)q9ATHKs?#:\P7 1'(w[_%Dy &g[N儎ٳV}2>}rf'1y.R6r-5U G:kI!5 Udܦs8T G ǟ%?h> D3.1*'b|#5islxrT˱mJ?Nb|r`UTR"AGc4*D=(ZM_R^ȓ5қo.rl)[&kolOs\ĤU `6=)()FD5.TB!XE%2:#&6&$NУ8c0) cPTqKb/-G!bX,," * o_gdܓ㸽CXto36 /4i9Qo~qa߱U=1(* (U3'hȅqPp9Sbh7Qf G(gh򠲩֯285.DhZL%@1+XGjQk^byj꬯_>4JdݻXK׳5!?^ѫw/\=x{76}.x:O`Q "3 !S8drˤDROTAm>HHa O N܂q<QE.VNs-V (*+1@CtbQ&wJNH3=br<73DbRMz\]Yv`OޭlMoE<3<!)hF67 q6iJZ$(8D(LtH c 'd:=$7N)|"OM,84hE5 Mr% ` Ik2H"'uGbD%dz_,QPI7D5eh3&Ҡ\(1%G"a[wa,uhѳf%CC(PTڲt Ćy?Bx k Q&]bPrIʠD=H--a,ڋ`<Ꮓx?hHyZhn#w^no%?ǫ,ļ`t3:'T~Thrk+ChُmmI/R#ɢLx]};]O3Nmߨy$9m{}(2gpz]90,]D.V,"|_Ϳ}G&%!&{=sLB{QW07qi'Ve{tI߅ uc}9n-)]7c:'7v;{Sa{GM3bH8C=\~>iyJI瓫&N nu2$I1^6]_ntV Q|]wʆBVl'|qyTo ܅!JX=qw w+f5VmӭmY7Q964?Zƍ@|%>" rֿ')U̪My !M87}1_uo.g2#RהN1+1_xfǗ'-e`f.LBgxn3DWZg;\os/]0fU=G&K°83@]@3l[]Y7Z61m֏lX-/Z .>w:YVLD 9oӍBs˾덆ͳv}.f}X6)!d6D. \CH"}vJB':xBf6sڵ0XjK |$$ t62*.)u w@.s`&a9Ч,ίN%R184{>eq˪g螡O?' 礟4pr8mޣ%©Գڎp\(pZ:hiU3cx-,qչs'`8-7w=|᣿YDJƀ^1%Hp%9OI0lB.%eIƀA*ICsN\T#g7/N) Lz?iwɻc5blg֘l`f%8@>HwЧۺ{M &e;k3W(YYޘxp(.G,::OqJ 8j5E֎Q wh(S#eM0%`Fwz[ ut&{^ U1ަ5]]m[K;a"pCȆE`ΫKӯKv~ʵxbw6i>3g=T|]n\y[b/5Frcar=]z5ܠ\Vz-~l}^ZX͸,յ0!Ԏ{Ukj״F;f'.O;A5ut.Wqh;̞eg;n#Ǭ/avwI♶Bp45(hK68!^@fQeʼnrKäH# '3kMrAMH(]P9{&Y Y\2%FՊ#?CcO: U_՘x ygZzר|]"t`Dj2E]I4Q6R%hjךCH/_uP,^PI\*2%O^hÐ6Šd@HFSqED !*Ղj$02nWi|źW3m4E)-x2jŒ3{B-%1FLFEB$Oc?pRM{FlXƘH9j $"P8XK UcUVz!$B=&.CUZ r-wmX,lއ`Mwo;htz0lDkYHr{1}e%Y+@lH,~Ȇv~}v㒲ZI.<m+',I.>i6\U8I]M6dfNʥ.)"?U)]GE2[&u$]d3@Hw`)"g#tA J1ugf2 7w69 }Weo%P0 ^_tb n H9zr&zF 鹫B3P8Cp r6;_ld FR]\OoF狏r Jq9KqN%/3Cx6q-S'a2[SU;+'Y\V+ N?V.|o̗ {>sK|n.` ~'~45C'XcO馩 ,.!,]HwDO}nzM̓#hs zm+%HY )|5a8W̒lXj*5ќ0fh/o%+߻ڞP%᜗KGQZ+p|Sޘ&x]K00T+o{8sG$~xo u^?{IЏ&G@#xyKï6ꮩbtꨫ ߥ_q_&ż_w CZr_}y3pe/ӟSu2$Jrkt-z`K'S#V*)R dZNi7ơNz?^ށ>tZ3v}OZn[+"N1D;a'$7TFa"j"N L"EZ #zPT9F$㑅H>HԔ1тFQ4`pH Q,FΎ/w'A F׽J;3=ZkrXγ ޫr,#LzL'F?'mpq$k֭fPt&PWXo`3n,>َ6-DcɮWmhwHɩ/TˬNN,){YF Vspe#:b0=L0.kh)iX ȕ FHCštJuuB靿N/g-9{& J̌T G dyQ(0,1G M;pr;1'` RFjR #-A(!QR J3̺ۿ8>!b [Z;v[fMs<2J#=R'aLq@8A.Qr ĒO;g8~hd=8UVb@)`OSd-64pS"3bOâکcAsgI(*&L[NI\,O88Bjv2/1?h.Y{$̦/62hس6CCۈ*PԶd׳ Sv'%O (|q6 QXi52A>"^:͙0 $D jڀ2F؉{}ooo 5ۃI͝$.2 d6w[h|Wd2{}2Jɦ0EٯC=".q)HIʧAs]f? &DpЙ~a;H* K*[:i%._`$lPm)ep3~7s!j%n8.kE{>%0IkŊ$h"ktzJRZp ?~SzIzyC uL7 5ԢohtX)ݼGN0骊Zx$+2ri]5, -hwܺkjL*tatUAVƪi)yƄ]ʡj cn&0 ҋ (U:E[ŲN֝'uR4-Z!6Zzh +=$|hxd.拇MtMux`*s@wYwQn\9Q^7WKO_xU| ʽ*~nߙSуG?Q;8 a^oqbzG-=j]WG%/8a_ DB#g7bG93_8Ǿ(]_r -y 1qV{[E NVF(AԲ넸Y<0§۾/[@GaM^BPwgYmwW%W_ σ1ѾN޸NN`od^Pj6 gy??,h 3K3wrlI`ZMbNN2}"ɔ<P;FӴ&üqFi/aN(Esfͭ$G~Zeuu ȵOG c*bTJi84IU:L,3əBSfiU@9DŽRn&x92J#"qGREbD hl[9ە \\~WC8Jzfc{yF,C>O}T 3 3ADV9  OAs߁ǀ v,Ď @Z(RZ 0akBBy8v d`tJ=#Z[>c+Y֚+L=?bA,x=X{1F`1|SRO  =y+-ݩ;FSܭ6-V9Efۥ%xӻ'oFa7AIƙ6s`s@.7 aTNyyeP?pOZڗlJ舢$^i8:i"iIb9JF ܀cC՘ha<˞Sś>T*-&^~/ս>7l)/6[/.>^ǫeTm @W&DQNrαYOpPL}0Uv0ajĊP%%TjL+#)P'<#)֝rJcVa x$ )[X1c2bnV5ͭajk<T~> {M bBqtoq|})CqX83[LH:f4]»7=M6mo8[Cݬ2j.V3 0Eu@T;/ki= ֹ]lv>hy^`u 57pt{j懃}/mϪ- {nU ,F]H[f=Ɨc=95TB#W! 狒['?l.X7'{oAzWR@(ɃIJ&I}86^bIދUz;ЇNKzs"I).(U&La'$7TFa"j"N L"EԷH}{B0*.V:M˽W3)YFsGNJQ޵#E/l;|C.3{`0 "YteO{lŢcIfKdW,V?EJSֺf Q#+PY9ꚜꡔ{ em03]Ώ$Z kusyWv橱PS=:suqfz{{N=XwHs]Qz4{MtO_|lv|nm܍`mN26Z3%JsWH.gi4m̼Qխ\ؘ4xLR"kx⌎>$ѣZL ?2ӕ%Mcll8[앜K:qxi噟M?wqvUm7,x]}Vel(hc4/Tk,dN8%:e"N{1$/:_Xku"p?},)@dW5z ⃁ұF"5UvvΎbp@HR!WvBZ:VYftwW7#ڻ~^[\xy}:e]1|nuK"g}>ڗp M!6dؓ=mFmFٍ\TKN{38 x:9_NӛN{]>^7=VêYꠥY|z|W~Q?rt< YR\ر_яgƳTΗ%zzxu_ؒ3NrQ^~dst|r4?+>QI}ãZԿq<7;V?oݫ뿯_zoo_˷_O^N)4v/; @|]*[v٦Om -yCޯqt{C 4HnovMo~~ icՑ#'IȂi=$V)<.a1f)@Cyv&-dחqY崸`;&Firq(~!" Fz,1@=/% : Ls'zTzs4eQ:ݘ@Eۇ2gْ*^rܩlZ; Brg$ٝ`WAyr.mS F≜6tc=k6-lv~7O .BЂɺ:j1/:kcXY3 dt-:1;7@KPoaX3E!d-P2 LZX(g`ZV3N'Ox5nXCưԇ)U6d)lY8F2hb8kL<Ͳ2*CCFOS&˫oiw”F+ 'ѧD]20_l_U-l&o lv $buҦhU "U>gNSWCXU}ЎfrW~NгY]XN6[U.Gv HM(D c*dm̷}qy3z \^_pO%O~+pTҏtt|rӀ1>Fbgw62s.¹):eGf^F[^pVGxl/g %"xzܶNC^ (||OjZF/a4w60*lF|%"qt+jZ0\;z&ylr Y?~: N;2tk1x.Zt`U IׄnIv{rXtSO>S(VʚUDs2@/hW*8Z{N(2 blB2"j='abAeёQSwVs2xD2)>6f͆E9k\I_eY 烾|.CEj_<_mxT\]Uqq94:c<ȯt“gI;=0iBvV4vEt{)(0]@;f>;)i,!ꃧS5DPW\RuNթowtqx:ޠGU+w4S.gO$v?{ߧuR%Aaüwdn f*MU_ %E݂.-WԍKRig{HzyH95uƅ":u ̥eQ䤌*=VVzaUPրnEY7dr򕁮ˋ퀴i]i/U}!5gRS{ HŻb'e( \tY&Hu?)>w{QEp@}hKZa\ <F#TrY\T4C0sӭM l7t@ʢ@MJB(Fb5IF| T1: ]Zl6۷6BwW,>mph]ް9j4͗7hj: B\b=Gb>y>;^bo/Ȯ]HG} ৢJ&IcBg*$%F!<?Q3^8gQS ʑfvT7%$As[*|;#-~*"eG뜁bh(6x VQ^Yfglu$b>qH'O?_ ȵe޺ rglU ==R?N[OiƈYTAb,6%8kTJ. Z@Id)$#W RP"i P"X|RJ甂d5[ƥ)fù]G'ǧeUW_6{t7 g7Y>V)ߜYwQǭ[| כek5^ԍllpCS|v3{!Ձ-\!w)FK$-> 57մ!2;LU[ )JhHi61r! Sd"7Tl $[ v ,x2N{Z|ՙY92d/K*BA1J6xX°ҐU,+%Zgfm 쏦:Edbua&`2HwT)ޠ-,S$v.y^//j cWZG.Ye%) +X zSY:Ծ^ͫCΧiб}ұ6g0N^rr.i+!2H _@:M`ȃ*NԵ \AHNh^MzO3==96nczmgTug<0l#čB7Sϖִ-VE,%%K`Go(&L,h !c:3f_7ֻsQZ\Tlzxi^.!*fٗbQj kJ-cl 4$c_llՈ곝Ud? t{|@O'㏓[lm0L2Qʺ,Pe`h") c(YI+ 6208$lIn|u%GݭmI-WWXT(Cakڱ-ma5LW\nm;y-M ,\+LlP~ޮ/+㭨 ?lZ-0 5}Z~gr9E+(~yS)/9W6\FEcbz_Ng.j؋74[K.UUE?A\yW#}<_\oi7+k!_8n?@lJ)<1ΔLʘSQo\hs|'p,χ<֌j:Yv֫Oh)O^4".!??//Em/Iwyow}V\FXV[7 )אTG_&{4Mg͐կCO 9xq$|SbO4\;@ :UpG֣J[ZRtA)g҉4by:Q֥m_=c"&-Ptږ9$.XiB4%u"ϕ5Mۼv2=o)Ư>Z]9mϩъ}/ӏS9(Ŋu~'K]!eDȽe>uIk2|BλXGejߕ80X0>frgřR JJ_ I F$2|HPRi CU! ;lDn9q6n)KFk4&sGR$-a&9m""Qet  F|2Ug#x@S ĩv!ă2LWxfEy:鉶(iؚ8:GOw8ӭ__ x7jwuDž>} ?#L>ڑ$, AzkZnq2.菙uv7ƪ &bEmPXl1w=NDȎNgH+iW9KPیt'rIAnNg`Zޗ{TbSUB鸗"R;VWt5O}##\{##%j\Ash䤸Tn+^׮*sIjk|n5BV?VVg2Vxr# 3e4\qAHS>ze*Uאy[Oj<^1RH࠰;m<8v(Cd(\2]!Z[gC=:Vz%V:ӊ}vWfw=?r5#%."7֣D F*|HCg,r!:9qJH94I]fׇ֧yY(ddJ <(r] oTገ'B\@ΎwBJrO4'ଛW&&ꏣV, aQsY,1D2:cYJ`>I?gM&5[bl2,$֑r\%&DS\ 4Djw"Uzة?O}y Q}d /bϮ9̣: ;E~>goZKEJ?&/Np:߿z{L>>q- $z\z+E0{ =Qy՜U57kؤV[t&6޿]C/m U~*}zǀ+lk9Ukb O.'JQ|)}tfT9L$PUh}6?^1Mv$:$xbD HW}Y?㓟ËPI;.1͉"4nNwcL^jZ gmG;2E²_vC2Mf'=i7z F!D@=7:8MAHQ0)YbEdhtPlBrݺc5!EBj˵bI4yaV3r ֵF3XKlӜhY!ZUd%0K08NTzJ&CL٪^:CxxK2_ݙ"nmr`= Ny 1)I!D 3! T5khA|WrRGΣqQ;CaLK)o6vs.c.v-/{9wR"!/wr>|(Q. P+5g b$˙6Zj҂Cbtw3;0fp7^h 417I,+ Q1H~ШZX2LhJȒ΍&nފ~aǒB1jK@Ng6d֪",V< DiGՌjtF?Ѷ`_1G l!YIcX\"i!$bdZ(I^Gʌ8Ne)6>)CC|? %],*̴4^Y8}&2'j (Fk{V,j ~Ћ_39mjNvb @Z;X9YeU(=IE# ϗ_VDǭqGJ:iȹa[}׋8_to '5d0LV{ Ґ;Kj]b u!ֈ# v+jJ#Ќ-wsNx'_YҫS5M_s2@/ɠW*8@=Ә$lR "S0 }%O51Lb$O]3qNhIBlz|ۉ{Ȭ" @gjVr/JNǑ;y{^3Ŭwi.vMƘlI Q4ubcb0$g;.}se1,r/^84*vv¨?Xa{@OeR1jCۏe6Y<2c7~Lorsl~;۳Ueߕ{dcd}|`36mIBmm;ڔZ=#\jk}@.#z!Je (凲~ǠZ?d $:uL }"aVƌ-کHwzu,kdYGYaI#ҕN^^lGRLkZV5-rY\ki[9UJ;}-rho\sgLkiuM ȻR'e {.t: JY|]J|hH3s@X'XGɈ*$Q0%TI!z1M#14TZ3n ZЅ@FZ2dDI_9UJW:uf` B14VC-Da,~0>A+!+k1z/[]jcK̺}1<>OЀ~N:dbQSTIH?)t#THBYX&L7ʨF:gL!-H]Ц6WhT^Q0I8'VJ1wW߿+W7k!bnvvݡxsR't;kngiSdUŋ0ڜrMt,I#+tB!ʅY9H{%HiC2m R|VJ)jF!dpiLg;ϗ/{ث=(gBe-8>Q>nz|߬7Xn/jsOYz^jg {D*(bm]0t1h{Jy]C/hcC16#W em4  H AueUYu(7mkz;};]ݛAv m2^ہ|.uS-g_Et3>B磳{nDKcuG>W ";%:;)M~Lg|2-iXʁX1J7^Y!_>u^˶q'-OIPV#NIAJu0:{W`tB[n$y

B|p@Lufبv-{5MQx@"샊(JZ0GoJNDh cF`@׾MzGױHZ*0b=VG)ŪsBhg3 iPЏXXx'Que{/ Bǃ7_ytq:N/7`ɌdPd`j"K0I'UL$I!!yoCu"*SRZUek5T@G HPL |@9i>-s(L;EÈ#=*j&D 6C{Ri&e gNVf dI 9y6)J^t1`٪j"HRQELH"S]\lLx;5<K b+8 OBDĻFDq_V˜ gɡ>hkwPF:|V9PS"m²UI[@I(80ev(m5vD'MV)NAk&qDXQ'E6!u6ӒsqQ6E9∋4^ *@ #L4}DfdAv%Da$8_%sša38Uca!*vXYA1IVJ.PY25dka_%.?Na Α$.yWBEI(8#s䘣aj3mF4@{M2cS+ Q%S0kEa*Lrz!!@oZjv`&?7Cnl=vTk/'8pQ=367<!GYG[\aDYI!1CabⰖ%Mƭ1̯q${gD<'kUXANJ&B"JtE:ZPHHΉy O1O~4B.[")[dDɥV8|EV_VFof;!cġcB})j]zAF|W^z >-SYq2ɥ (h$)\)( [J!?n0~NhZ 4M^E3㷊m2ISXn2O&tƮh ?O m=X?k爲*fV+_.>1܉tz_I۹^R %l4{*=YbcWcֵs.}dZh~S1;A@TnN4b1Y>Mfe# se_LN _忰I4fGӲXy \UH8=6k2?pcL>>ks;@DkGtkylqd·cJ7 pI|h=EZ'mܠ}$!n2zƺ6,VjWYձ/Nq'fiɲW!Zk6۬f夋}e}_#n=YI9P"_@FD*4>"e(:P540װfiRCjҬw<îK!J/Q5Ag#5զ@WY҉aЫC(ˡ`7Ӎsm U8aTO_(OBUZ ~^\@ 6j\Jn.(s"o'P $| >xVUݼ}oF[b}w-وo)mIT<-0~q0rW&]SlQt{<|'$$zX+7QNCazJ5`Av5 3K^ȳNDPp9|bOch_$אG,W/ |;K5J.S.Jdp@VN~BYOmYgzׯ]BR  2p:}-tlV*Ka9e~Su|Y\ŝW%.XVnR123OFӊ|@I5(58QR4SFWg Xcy<^M˓d!}gP,2+|p5C8MhgŔj%kM7̭82M7+9.M7CqhJaEӢM^iU`X2)=ƟVsr7¥>x>gPPm\(r{q'cƮ zi.rB!.z(f{oqhclדAoɐ34zame XTruwBrBiE­~ wDL%/*樵ajyAm[MP-mz_VU3M@?I8jzZ7k{Y#؇S5Cˏt`3eR5i^NvDw`|mh ̒rs "mBnP.s-, weֱYc"dòX>'zV9}+]oY֘x8,o f^`. !_qQ$LpK Jϖu~fo#>ce*3yJ. q¦O=Ę6"#L2nti+Ճ[Q[K]~{ s?ş՗_p |C(BB_%jK{3\%>R| ~#\k(l?.g{3t׭>s˹dŇ?Jٸnc7MEQOqJNEvy>Xݫ>YƊ\cn"z~8f'@ynRUz-H 5kzT}\@\Syc :Z,h<+b< f.3*?%?n`!nҠ.#§\ ߳h>[o.Bm$4χKYA|rsF I7 W<:ϒdm(ݿG#_nގ>7z;abY-;UBTwkӪ.,ozBny.ls|rz<_au @f>R-PӐoN*y2wcKSx j923(Gê`f\9ȿmVlhS`/_mTyiZņ"h,F!z;Rxr?rJq'Oﯶq!؛] !,}>[o}4~e~ ‡.fi2}F[-8 ?b1ϞQ2͆@ i}$_\@*K4a\g9Z4H+|{ĤLh__TĴ=ڜRe.Ru!ފWV>_TWU֗ͮ]DWp8Lt՝mmV7ׇT0:cYJ-R!ImʭTIi;Iuz+jҍt٪<\l z%v;)М 1P™\ s)9sRBp$4NL HYHd~SY)󸔔q RfBJ3ӌZcL^,i /Vwd8= :`{$UM9RT䎢Z={KPMZ)T˃dw^Cޯ[)5PgYv9RUg<Բ3eURSY*?+5:׺ ה閦oU`֕Ҽ+>:y$%ч̻t'緮/}Q' > aSaihʊ@1.tDggLfYI2bug0nydH(r{y2L R\0-d!B) q"SeV1'XpjKN30ohZBYWq` Fs ٻ6r$W{@OR|>.>pۻ/ H+đ=lObKe-)20c[bWwW=|,>b Yʦ"lxeHjp@'2$t|E M$ q43a/))v~BdX&fo06G&鮄,L*CĎJR'Rlŷ_z $! %CImMdHI kUAt@R+Di'Ռj'My7S<-rl'C~`rQ5 FI"FVf m9䅂 8Ǣ,/ *^O }x&jU^m@?i6π馥6;t tƫԫ3̽p2Lj]7cmWtc f.rl:d pD3BKV6' '{Xs8Qu ?qowd<谍׋ڱx&͟nOS)B'\nS0ƽLmᏍ8:St!wƼ[؝Fh u!֊ vB& 稪o#Gsgգ< iOQ Rكs2`6&@Tp$V)< =1)ThX} {퐃Oz5  !SۏLLe Ď^9[IT^7]U ?ϺUw Xld&בB2*~H1K1 'V#*SR"K%򑶜@ҹCd*QT!-$h(e12$ߞ\)Uu>yJR'1eH GRH)Zjј063e;!{4.G~G93ȠJZvLl5I"DNk@PLtEHL&B@DIZ 7%,sA6v ufʜ]ii_C>f~r6+i1t2|y,u`kO.uz|p3K_,6tcO6P^}PEtYIBk衤$ #Կ"ZwH) _ GR1<Ƃ+hQJ*Ff֌4f1, ,xpu?Qǧ._ܞ-9b+op(PtN205%@ऊUF6.4hg 1d/:%UMm],Sqn'| @ |ݸZṞv38CcjO vǫjDCDlN'/8ե8J)V8:Ym][%cnd9y)ΐJ,:klUQU5)( &$"sR]\l]̜x'U K b+8 BDĻ&DqhmV˜YrO2H'*qQ-Bŭb(80eN(mULUF$drkfG_ub\\tXg3/9ec\.N0`2'HA9obD`'*{'dWB 5[?'\<.fF+8Ucùuc9,1'=iRj?kO Yw̻sɹY?\~L_(3UfSy\~G*.DI.;OSk?ȑvoC _,+{T jIiSV $Vh~L \)j"%"N"$Q Hti-ɦ^؜ ;1twgivv}‚r9o7pϛWG5߰H0EBĢ*SdH*@t)㍩&Dh՗Ƿ? ?08G:gLEpɻ-ڐ4Ѻ[h"kgrZ42\!*8ʀie88&*$] f(2L=6̜-1!@w/o}7 fw] uU3v7S}">ڭ/736}9H ~cW89)qxI Ȑ01qH@p6sq| y7>YRƂ :V4i6A-DXShFb/sbB&Sz=b x^d $;QrqRq. 8|EZ+9/DIf+NS7)!CG P>~F|^v,s2ɥ (h` B)\)(孎B9_ G#`݇faU~xxUzo9.`'b:*f_8>mٟ/zϳ?ξ,^zpcLޔy5Bjx;k۸b*;kgiEr!x8޾wkAWF5xh!hs=솚412zfwrJ?W徘^EE'!{Uo,as??bٮU\'Ͻn%e毵@y$ߙ~(0(0~?tD3)nz:_j"pH+\\\-O 0_h۱3NWO[]L]N^`&ᖞ[=m>k{QkÂwb+7+IϚXO;.Iji;ȲW#ZkL6ǎ۬ܦ礳=34M룉z*&f%A)+sDQ0M|D:(:PDo&1 /D}}+jwysed]t BY]V]:p(>A'D #aĘ7dx{)[Nי$9/U -7ڑBJBG8]/$N`T'фS&hCʾNag~MX$(pBh_vdN3(X6@$y8hgźίhYw?PA=h1C)w})箛\~ywmIt7YGņ,iV{e:%p851})6 !ْnT|YGfYbOu)~JC(TDjgj{ /K?q4ޓNl/CEZVr6s8&gpi&4m|'T"ī(pm|CکmN\h4z=2FiYm\D"Fӥ)v:|gptrG[0\l_/f$=_Ԋ=?]}l30{:cv{Ws!^M;l+% |EuףD F*ϷsO dRV's I1rΡI/:~0wl+f&рd #SJAiT^ IhEB!JYrw@~XmkU: X/,"L;$>NYcTщetR#0@Qr9iiE*cP(" hxfVyqG)'rl'8pIppl֒%olD )Op.3K쐞?Vq8P~;ِnTQO2|Ź%g*&:==O?*9 țc~+!4Sy +>B~0P+mj.2#X-V?y]=x~1|b8 <{}X-7]Z(BQ|1rh3 &I&74hWdk56QQ؃u<F嬣':{1r{FVΈZ'y#t ܁/&_+'z/B)YXeqb9_|0M|]RKW( 0ah4Թ_{7E|ͮ5*oiDI}W77wՋ_?>?Ǜ_=2}go_?ǹ;i9a+5߶_~՜joV5װN;mukvzz^PrQB4;e#O=Kwkw[]=qGzsW$cЎ'C=(A^jZ gQz|٧U(Ӟ_zDk"p`lhGL 8|\^'lEĆ9tSFi7ͮr빦kng !̽]=^;24$GJM.qfKE#6.!C_B,|]œ@]d*hbޕ0.1dT>0֥C\ݜMGNz;<7>;p<'틗}:}>}FJ@Ņ(4::K@k5hRwKT LIFj7Xo߂NIӚj+wIWʭ-yF@7MhܙgiY}`N"Cɿ9P AQ糣 F~QN^$~sƗ[;_F(~).q;}r;>]%O2 DV%p#VzVMW]g=:"~g8m҃nի.Ur>[nqvmzYyN_ꏪo?&ɯᬂ@͞cb<{/~ySSE}ȶHH{9my ]P#E;῎ *Ηc* /O۲Eֻf$T)TzbJǕ/sp+/s_N/'ݗۨFFXv($ $m+#<]X)J@PC"*S$LZ | o<QǸָD#DK VH ;Nbɯ$KQ!gy!и-ɃF0K8FEbyUv{H5#ծE;0SMDAR1mJ&DSR'qѨfS5SL{35Q͸V TnLH!.v9^ўSRS3Uhc"1Mhp@ :g(@,J獠'crfytoTE>_Vg[8_\ŘFe5 \/X%[y*3[4f^*1xc?x &?̕6B<}ғcmIβ\/mbko~qe.{[~U]i0n)otuNFא0w`B\uEAayC.{sќ5)sN?ǟIC ;d|s;k"m}s@: ѕVo}gZַ!ZI!B( UFUF @W ]5^ʨNWw `KWwC+P=;X@W@WKy֖ϗ3ԁ]Q@ꕀ`alc>ke$”'꫆q`DtlrBU"tXL>f?}&kͺEߨprQ޼|v$r0ai)<1Δ g/9M)(%+#֢3VFkLW Dw+#`eUF{߇;=Eʮ ;;.ЮUF+ľUFt#@WO 7Tt2`g2ҙsW/go̚Co UyCK*|Ϊ)\wh7Ppx9G*V?y4_4m@@R]F =a8L{~2OV > m#$Bt|O[KqZD 216mÁ-xU 9ՠ-E@IP&sJFƆTdH6Njdp{6J JP\Xd%pJdXz0bAք o3|8-IK-A[QjdA%'Y!s%+>JFԾ( (ԨfcAcq|G$,jd_LG*VwgP*aU|G?./(GEq9cy; MxWտy]\Ú8fiI/F0nZ<+Pi1r~Z|Րq(Cξ1l$M #D&8U<?*):w-GHTVՊߪsĉDoz܋XVz,JK`b/n23]|6)boXe'?jXҳq|rIٮ5Uە5vtUzfC5c 5ͣ\jbe컩Q.TLcq K:CW.]v*ܷܶzA$@*6YCO -NW%Wzt%֜t2`*՝YG}]=*CC -U*,&ş~ZѼ\+ּ7#ˤӚU\OEiӜi~HD5!ynx7r2H*o_ @H3eKl5g&i]k)۲^QlP^^.ѫkJ.*dNEݚ Eq6Rl^7qìږӗRQ@@"YctK !l&5X77#wͷ9S  Q]ȜmytD@'-L8ƥJPJDDTp l$$:Qmk*A TfĘ(hc&γC(x$jH̟maDSf0űZLB1B Ta%҆H|`p`,%*dNW@P918ШUېVɛ,(m~FtLY<&D)D"V[@XiDb`C8=[6h -M^RdmJC4 e D)F) fTH1)!(8xK6IhDK:Z- u}%zCE޵FFC)ʎ6r:᭶8pҷ׌sh'Xa=ac!s@#̡  "ٻVLU:I=RvyA1C'[ZpUTiXYi)@fT `6+(\xH +Q5 7!:Ci*dL'BXp6Lv:@@}XE[PB]Q[=ಬR uWh%Wk@ e º3 Da( VCH(PED&TDE"3|5gAx;b΂GA' sG eL1R"%8TR 3k'T b7XS9(f̤cB8*VZ hM e*3[!(@s!^ B((Ez@ߑPIWm (ĩ(HvU]T;jF J H/[.Ѱgd0[Qзz $$dA*2z 6A j"b C$CuH*мGwU+czh2&236# _ݠbF\*"u44U ULl^vNR'D̿h!`V9t& ezM[ZK ~7kW f=fU&>Hh"X 3 o b9PT8xi7olJ?l:@GVZLAG=$]I4T*#d`&SQdn֣2:#.3(Z R|$ LjyU,C Y0ncP&>yt_V 5ȤuՅ:@vdm=U?jPJ4%wV4"CO&XVXktdDPHc ѣ.O9mC^AJ./sgmG]I@(ʠvPJ4K q[O κ$!H;V@@]BZx _3.D[&$˽e;VE̓Aa8NM2Ӣ td!.h]`mB$1SND] C VA;8*jV|,**BȲ$P>flq)ت(jS,ԚϽ7h=V A{H$Q5$YI$e(m@VӥKoU4^";BZFwrAVmy6,a* 0dkR@n㠝{ ns>貸گ`Y˴m8Qe,j0fU:9 6IҢGJ${hwO(8w%[Y:Zm5EZSRP'ՓFCoׄ bL@&rh' <%2`Crh[SP.O(7"fh8(>D+YLw]zʠUbF =>A( =`-f=*-+P!>Ǯ;"($'Mm2X4kng!g0E?"oQ^1bp0) ,*1"@jt01R`\t%i#*`hAgМۦwZ,nFW -jf=ؤ=Cɗ Uk&dd, 6#TK.dP?uN^Y%aנB]h5х7[A[F}E DKwVT@9 Z6i/zBL M F:`G򠵧BSp*qKچkCWQ[1xx@A4D ;h̦rj4D X%d,tR@Ւ0[ YZPFU~B:CWW4"坩y0B.8! R G^Yrp.te0 &aXJXFq> jVqHN q KmE7o<{ L|bl-)&RA.nPtJ;i|dݤ3/emf 0 tts\L/8\*&G^Ds:ж :r>ݫvu_oߕbbjvs۲m9yh㪭dm[ 3U;0(J^pU`sWFt[kF&z/ tN } N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@:]'G&33<"`N Z93v1h`RtCL4kQg] ?F!F RZF!pr5 =͎iI?Ptfv"1] ]ic;?`F ^c+B ҕ (܈ c]ѳ ]ZLW'HW6 w\BWABiN\F$:c+B^]kWHW>:È ϕAȏsNWG4c2H ]}]ZytE(R9L}Ǣ}+!·\Ci~(<.R{ЕbࡇTqx'%@E0LG-ln\y@zݫ%xG+JHeHaQٱ B^eU b,t^=]c g+-qDt'+kXІ+LW'HWFj-̈?"6?z"OXW(wgoO&R#X'Gyvu녁a_tanǟ Wwy$\]<"# L-WBV_VZ_v5|z) W/)ObH=>YZLϞٳg8˄LՑ˺~To) *=$SV]Dzߛ/wg<+DCo]D&/v^W3w_N[Nhu[V'7<6&(om}Jۜw. :ChkW ~%iz|ӛW[l۝y훯6g*W<ub)gL1Z*WUuE\&e5!Үܤ L~^ӥwI6|^^̙0:t?ϓB[+L*:VE6cFn:٢B(] olӸ?K[!{|v/CuX#G^I]}׳$݁i`|Fj>~{;^Ёo<;`Ñ?plkOG g{XnL? Q}«tu<_%8ՕR)t*["xtpUaP?o=\~@D[2 $SOYud^ՋMj,i:%QKU³gȑ"/{L|`pegoY,0AG1,KIWll-9yXME6}bqH1۳4Du-̫x^"Ut1 ǴDqaoUMד9&jmqͷo:_4soV?\p%{|gú4j VL3 `P ;/1iohrG|SgmޤdL++UJjQA'Fփ#l9^{x095!}|_1Gl;,#AH 6QGpL%NIJD3'8u"zDCLAqREf)9̉`laؘ8` k)?0Gv8n?߭BGU oýz h0͵??FƷ ^UEPgA33>pRIuHgrhAIvW]jAb.o;D_JWһT(ahRɊJVzxW>۹;敒h0Q:zLoy_SY{*e%!?1B,MyEwW$2z$7qE7pƅ9|ԴԪw|gVbfxniOiu[[؃0V>[ K_en-7?77+|N?WphЫ=*1J*M8)B ̕2U}Oۖ?{o8?eNb'a 7d<[(9*0Gc6τˁId`ʲgFٻP"D+U:+̒CHLI%L@9$f"O"{cGj8|=gAsk?¼?y1/_4oث:)rrܗl'&Y{fך;_SKO,UbnU&,fx*|(RhR$)HA: \+LQi'lcW]le.ڱ)oB{XjjL)VWLo+3],5C=8U3V4]X84<5dLx*xk+)UAFSYf 8Bm<)}RA=VʥD%"ZD;\k;+o:wwv՗a?վՅy. T86z EyH3,sf$W:} }wy|;"Z-sTf*S::[h.`ώOBc=h~>N)@U5e=dB VEM%+1)C|!Š3 Η$%bVG%GU;۟4Hhs&D@$ OX @92A Z "xu "V*RTcu]u~>CX9Yd %YzKI4t3&7^G4uB%KSQ/@g= Ŏ }OYɁ;.wu[-Šfc>yGH+I?l,xY e כTf_IT4]$S:wdc9g\2q6dgxc ۳3jl/F*F/GaH؍Tdd6+\D-nƏ,~}sa1:i!xz!f+ћ7瓫Giqs,Tm_tm^ K(5=k?t4+ݙ/یc]铟o'ޭoJ8_/~u?]LVخlTz]wTmiLnlI-]mkFlofwc3=)b^h(K>n=^ٽNvuNn`[wr[*^Z:.HiX`TӼ;eG` T)guݸXqj~kwjٞxي_%IO4Jh0Z$d*]]Rqf?];; %w?|\wߒ- F2Wg[Ep@GMÛy!M7ـҮk[v5vU 0 m_ζ纏c";9:d"xZ}DGe+$#r=qbP@ h>&:xq.U>_Oi̔'-t ,Fs^LΜf٣!Z QIF\S XOdc'GNߥ?n9j)Nγm973ָ;.qMpUt&e [{)?>i¬D*`*E Rն?t:go L=W0 Գ%P.!r I?' Z,^h㄀d$B`.ѨMLl:iEhōۗ;:},.;X"S 5j1hْ!ZNG[R_WR$CbYVyQztwg7#Y-|H^b?i|cmKl$/‹σ})lZ(UymJz_ 4%%W ʛڝ#tmךԴ9KO>gi@eHL6>€ȴ +5h5˔F.{(i r"$xR'BFqZ0LFymc絳u;Wߌҟ;ʕǻ6RI9&ܖpYs#8Md3}fJWALKD%k  +?Mr(VZQ\9P)t|;Ŷ%dY6Ì,FPb4pn|P pV')A'+ST$G#f!Fa!%Pˑ@Б9d,Z%18 $ST1INn  4 vڗ/Kr|y*Эڒz#Ctst^آ##̈'CG?ii/T,^kR*j2I.YMh@Q)REFU@|rBy݂ǀ$R2J΅@xDRgqNZ6vSb;K^ޠ,ss΄xC:Y6 rwnwl{wep^dtBQGa7;YC:K>(htkvlkP+{TK58ӔHJ;N/ɉ&p`uURP)L|U 9ET{Qw6Nŷ^.j_-jQ^{ڵzVm `#OJΆ^9)>y H3 BŹ˕:.%}[HFQh'\y1(kR}P pa ! H $!5ڔZs!%yL,뒷:7YeEEk,UߵI+tBgOb}9*7 og!i[-<ĴX'| 3; "zO7R7?&Bŕrr|k_vȵyަZדi.l^˛OSU 2L11Ja5&p-="-N\&CbL2#Yjpt\+:΍s{o Q@FϭBG#iJ51qkFh0mh^~nhw-f;xE,, X4W.h}.k^#QTɨ+pg+JKXU'ĻdC뻻> dȬj F6rqhHWEf%T'وsY1q"ou'>&{uDI2rܘ(,#%%lhHF,+(/TB:X&K7g!)U<23@*QE +l eV)er0R 1愰#"ٍ9M4&~ZS"\LqyBwgNjSBb/93,sib.&m3OހYae/8=ZKmA b4 4ٻ6n$U{hURӮ[LjIJsヘᐢdDEK`zQ^A <"d-Q" gV\zMvX7g$VHaqzYN@Hl;51b.3&$^s/vdg"u,6;rK&MgGAt4(SIC&qFh-MsOҦTNrR!% @E@%2C}LLV#g˃wBѸsׂ})օ<4,HH!04djR 0{Ų,sǼ"c=S*"%L\b! N0yyL*X!tNq40kEfYWu4+fNIqZOv4N7ι&Ǯ؀Hzcm"ē.}\+;U;Xny*oXLсȼg4ZJ!ZB%%O[D_]~ٌC`li;YrACF20C 7sI9[ Tjq,XF,<()*߶(fy~W< Foe֟M?_8b g@J(BL1p Մ'P>̕CUVRJWބ# hQ046B+m&ێ9BAR`rLEjכE01qǮ QF5;9^P$Zyȵ:928FVL" dUFã)Ǻx>G"E2i\e5X4"a,(f,Q\3Kޓ)Gc'D'Du#"8gYZy 5٢.poEJ.$4t."㌕$M"rr Lɠ4ɔV.\ [-WQX;a5rGoMu \p슋2.G\3FБ $֙rҠ%SLyE)>^[[mkqx4cW<P솇u G[#:sObHF,rHяY)>Ī4/=ˁgY*u5-akD]Jh9="ўp>8{;|M|Wz$ŤSbzq8i~(57j^W20{|[r:|vzƒKMG2a0I%|>=K+[7*z u:(оwƁ.*m**2hfLhbl}jK|UvQZwdR:/3(Q)<0L̵")`n2?\K_ssiosJFO:v(RS_ӋoӤYPH~5SO~4ןA}q1oe]7^=sXJzׯ 2:{% t7VMn~TIճM/'Fz 3z92>T3 BR1-oReIF_WyDG[tK?1d@)\ȜEPK1&e΂9Kk.zEEU`ඤ!ʦ̐`\hS U(ES7fpLGr[jO.Ƿ[|) vP)>LmQˁ\s/x22[yMdlʊC ٔG<- !t"&$1 YnR jsE2tLZ 80[F䩂H6vC]+w:FSVg/Xr7X{vLHXrRǹm$6 IKX=i~uS+ٌU\ֶwQ%6 |W4Z[;͏;7vsټTzZY_~Jti롘?l#ҀZ7[.zw%9brz9)yoLBoxnAQonwj66N/7!kox`el:ԷI߳vXֆM2G^1;^2v]N/fkRgKYӫ ~BpUcBO ;\*pe4iewbFT`Mm *kN\P &o5bNm|ĿRwߝ;h:!v@uve8 u9@bCJ2piOs d!geGLj& ԘQj,3!J&5٧ą`(+D.@ hgOW#gLk4=<=JNcK R?8'}W\/wFt*sw lDB": Cf@xt6ʜ$zA₈hBs<%X4AFO( c9R{йŐ, yt3 uc2[c"Y`TM$4%.qd7n'W./;կ7%i=,εػ6rWU=!x9o[}k9/;5Ed#elٞLu-Ldu `۾☃Od9hщaWuRR2B!+K, P|H˫L1o'8,L#@u&59@SuUx"H;`W{=ҕf(_}?kd?Ӑu8WbG::/ƎHy~K&7w7\RGY=ُagtC^+_G஀ͦçN,gtZؕ/](Rk?UeW8'}O|Oe2}Gt}_K󌒗~Z͝ gmgG]J'JJl+i&p8*"nP'sZIqUNUi~~cNnߏTOwMN>>vG̿xwWr 6¬iug2Ir5 ibs, `"3n]FsT<0OΏӑ3BG&FtM!R"fNΉsdхVߪt$U5r`s{Dhc7LLUup|g2PyrD(iL]8OgbBzL(R-¸ZpNs B)R50LFVfU0ȵ\vaZ+V|&dѺjK5 ,jWduI|IU8Ga8Zƾ*NbQ QPsEZcH>eхb4kѕSQ5 RDmR&e {S9 IZM"',EȦ| 7-c[Gfb>d bцP-EUkJ5©:Lݓ4.)|6ɓ 1vM3b6 ڪB_)ΖtE.\&MR^MmjGAS8u<=(fָ[nӢLUdl;ޥ3%Z(%x+P)JTuQ9+{4 Trndb@ML`]V)gd{eBژ#""yyc1n*rƷ݂?$yK:fo :WoXױ*sj>j4LEi w426ŗcnrub-9-asmeMIYMHM^)Ę` UTo#LE&n8"mT0MJ*Ŋx+< !Ka=?ßx!q W*@lqZ0=e̹Dj/K`Ȏ%5HY^b wN%Վk_ {֤l1*esbl#Bk gzXJ;I@b)L qNDNZ*LeHR4):mﳳsMb< ٣}1XOtʰ` eqFkEN-]!sġetEF2W"ZN\ y\#iNC6aa$gbu9fH7 iX0LXXx#Qmƌ/*=3r"o C:;[\}XV̤maS-4Q hXsVjK-o 44g/VH`xޒ:2VTUXl52&z4bbw\̱v78cgԎjO QיE|v%D4xdІ\ 7!vRyLx MA.*{]t0FZ0JHձTL3d4V#A^q"JDě&Dq_7C&}EFG- P:GZ-^_Dj'b%յ 1eq(}Uȹh͚rچy?y$x1lkLnZr,.Bg\ '\|4SstY1jqeMVJ2yK焋/ɣ8<| sTŲ|^66)w YQ<7%kR ƺ~>MmkiIUt YBB?fvE3SK={ViίxWM*x3l6ك^x_oxz/yr\>dvmXyˬCï}ora{QsohꝊC9"c}%-G5iAh.*kR\fs Cw#ɉ5}dn֟C<¼ u݅>Y/gelҍ[vٺ8Oǁ.WmC[UQv-F K<K/2o$_?f^'O[t1l\:nٰ)B1rflT1_ť஠CݞNp+2k"f|"y-kU)= 39\w`t1)&ma0B΃3Z6[<|/۽s'6Klt )k"L%SVSKqmQ9J,jJ/~eq;ZFTسp[8H{67\>~Yn{~0wwC{G_/;o=`~va[1 Š:FPZ(bAgWz/NYHa :!Bu b5)sd_QsV)`*䱝TY ! ]N9Y]UMKU Ukh'iwI 5 L=M60iLG=Gj=&?Сb 'հCx|f% Q=g~5lO]!*UYnlJ`DQjqx" Cb\}K*(x>8_~|셞 y~41הW1kSp6ٱQ  SqETtAFS>kzAǻWFz]A`} g# UA/8#s1qK+ehnVX~ާ{lPу842PEz{vi6!d'ěonpb?TOHQh`W•J m4$b !9S6g~:^gۙ3\{CM`Q'mxp~:rWwWO],y߮\{_z(^݊r5oBukbSV3ϔhot}gre_ 4ZӳURL\_rČFL.5]-݌-1Y5۾67mU^Z?z#߇40/JE539=/WlWMs>kxBI<=ߚ\MyXwl\|[tm.sP6q3P˘[ghA?nfj%3w~¬2V#><g8~ ѧ5E*$^ws#ȡ1 ivW몚:>P~+΄~sP\ʷ C?ϜcekL.a͋yג/W+7 k|c$FJ{];7D-fGK=f;1a$$NSY~&{U_ {@M_0>완Jl:讠RD͆5}^or iIRt,Pj^g1oc&1QeXՊ|5`ŝNFhîqlkV9cYN/ j} 8!;mTcHϔQP:8bd NVC>! ҬG<]9.j"1j`zNCkRGekmωZ8 _HEK'$L.%chkgcZ@ŵ+4~[wIpmZ}ݱN]`|M!$>JS*SU^b$h(3I}9# $*Ob+B RdZNiz6BFNWZedϧIa |QP&ʊb~gqɿ uTϾ4WD$Ueܳqhm V.lY`ټ\w6A³yg?~RfeՓ /FO |gz5NJ=g?ؤe`/8q7^1<H\g~_g~Wu*8a%^ ݥ @% c#ЇeZQq [SEK*0h5BuJ&DVzp@o&r/Nr,#LzL''ADp( @[͌ۿ5$J [e#6mV-'SڸVZX={|׼쬧<=U)s-u"H(Ajs/S2BalE^]F(ݡ!y50;M+{XKv vdp4Q:0Y0ЄtI-!'njh!ɳU (W z&y$,ViǬQFYJy>%\4]F,FΆ C orZa'[~CxNd[huQ*'0ςX :y!u)153pQJ@9N#$bρH7:|0Snj8(ŤDJNcRu,p̰4N+!xz +?*9a(f툇g֍jJ$pnA"RQ<2VBD$_JF\dtO>մ #p( 8b@VS8H04*-Rq%րPaU]F&m(0QH"WJLᦗ ^ O1HaWx# wq v7OܵB?ldgo,.=ғ+\Udm$W"0$7Qb%?Γ fVɥ~R؞"🿑ؾ&OS<2_zX2S`)"K{/L…|M0y[ ,85d7Po1DA3,H߽+EQ0)x;[GƧE!?٥)Z%X'oUσDr Jq9K/znV0lצF{+lj .oksZgedf_ s/哗ŃWϚmbrSʵz C᫫>lqa!d$ґ_<ŲaaX:LpT``YJp^|EjrtZ=j=e.+%H٨#=K>y x2ϱ Hn>W2XO2Wuq]K0]1U'M9|`w/="wO_~zg^| w_`_u[Jޙ_v"5?{54WMdVg]-&6doˤ.|! dʸsjJVG(ag#Њ"Q5RNh@; q^In9)Jmq%/!Y$+<(1s"lUi <0@d}TZC;9᭝-sSn;_l9RN~M 䪓0ZK퍋"g4`+$T@8-%&jKWܟbovɿ+6osb.E06 {镓!w^*EkvSIv|ǣ|tT63pbs}{O/K 8rYĔ6 r-,K`_2.y\EN VSMaZFlrA$G _b+ޟoUCIK֙t(\M՘z:D%t(G3\U"Wt $XJ0FP*,lȥPWZJ2!hD杁+ WHW}U\!\);W@0g3p+p=|eJs%\DU"Wu&8PKgOT+ֳ"7̦#mvtu*L<{xZE6eszhW?ٛ_}GӲT9y!A$Ku{ys4p黑ݨUa`NTt`J0LLoR4__M2B }983PFf`J/*4 Ka ԎE\\qXt(L91)@ѻ4fu1EfXďf)Nb;hm6ؿ*z+|Ǩ}K]3sY̙C:<NcʠG.C-~T[m"ȮܶÿmWxn["0ŸCp݁D=xwWZMJNpupEK X2Bn>ݨdWGWLHh Vu3`VCD':BpS!J(@.#D<):Fq,:WuZUc+)]`IW\Jj|pD*Opup@dc8CDra523SpWѸfX"-RMLq\LI).\"ͼ8O9qsm˃wTҨβo _ges?Nqz)~ȋHV6QGE!5!Xh-yF:H*0|揲]pz;V պXFS$ZeBǛFsǞ.lBA J\{ l Pp~Rb]]!{dP92JF 0.S2i:| aPժpQ`'U4GXHH~`-/猒ɣ^Dk?un>A⁍0>5~gVV鴑,zAz3?WL̖jF˦B6OG~Mt ƚ!(=] /ܕϲm-e#=^sU`}\w6W}o&}bEhOgl 7~&V-WrB:͏  f_m}y9exӞXL-.bjF-r|ג)E 0%cUCṈ|:Ćɘ+@7Fw{;`0|LrJ͙"<60;˜̗xHnO^ذMrÚ#n.FsKQ_%7)-n<Msg+ğ@$4цcK?*yzhspmX^#_å=*5-Y|N5?4]W >]Vtw÷c,+u RզLK\+6.ۡZneW=|k,f5N\_JFZVF@E@! e&-Q *pN5+KM] ^uB6tutřp9X YB6th:]J1Е`?jwz(]`Dm JYB.g@Hc #]IΙ!5+ ]!COTw%Q:FR\zU#B>vš Օ]!J1ҕ: hWV2ZuBU;LG+#Pur`Um ڸP**:BB D+D[}4/DWz˦m8[1V*2N6S0a &9nšJvX :1ZZxށuCnz6+[)%iVT Sl04g}]ڍdOIbiŔJb9-B%%FAշQe6 hlGQֲ6tp ]Z*OWR$+d~3;Q5 '_|Ja6?)0a8Ryy„EI_f_ G]}Օ>Ija]KIa=t.-Mv ΖcuƋ>i `u[C9霬}d>&)>H+Kl ҋIn)Npeiq?J9*eO ZͫOyt {دǽiYoBTҿYԀE (Ki=_Mږqk u?; _ Qfa_̖MopX9Jb(- qZNe* tS"ECoG2vC{MvCi+dv2Leܿa^YkB$φN.$RC}ʕޒr896>qD ˌ0o$xJjDPzv+x]HJ]uE7${$ˌ`׈0?.thYuBDY φѼWƄET ]!] PV΃? %wAorbkKY>+ӧ's7h'K?|]=|oK:=4yf΄I.r9>;~?; ٝ/SN=bwy:[\b?N2R G|gӫ@ܟgeJ@wt]>)e7B8)s$)nzHpn} pKs2/魲\88{g16s8={;n;0Ed:xSǤr))\KRzkksAU.r&#ZcLș349 fmReBWNjvNA&XFT! Q곓\Wgh/=)VE /-cNgcn-x6MW:ΚON6>R? 7*1@K:O?0{QRP(&S{!tj1RQ/`sd6ʲAnׁZqk8b}5Ot8YAE&"/ӋT}9b*$WЈt:EQ~hjQyZ)%Lhڟvƽb5$^j^KPyEaL%m0;1ˡ~mɘxC Te9{R} կYi>T1>OsL3σP4ݬlT ^J3~]T~A|[MPkpB/MKaY{lP TsYs]W =9ޭkR 1f= .}kYp?E6IvBbMBY[^꜐c;x|cWE(4LjM]J1Pq{cAveR11i5eGmi-bϊܭ^ j5G" Vrznq;,&Nz,WRt3|I'z}Q$\tdDjy@*` \>(Uw9PI Y YpӴkR]͹@*1 ̓І6 N``TCH:BR`ABmJVܮ(i0ߩp[t֕κCk%i4KR:$e҄o cyHJI'/sڔ D1޼HIGb-Ej`+rƓJ9 {m$A3@ *ߎT9-6,ƘD9uehEn#Gc\t uIĒ> 6ouTbSkx1g~,]-Lw`͡w ="yb>:iIZ_sw#>a2I}0MN{*^ހgUBq"޵dhS:r#甇Lҙ_'S vl~o(] /SI% ƿp-c *ϗhG)Jrgh-k燡)1!t>D2%٧o &v;ӳQgRATδ?i,,oSjD+Xm~Xq CHB*֕2͟n;7ͦ W`ns>yᴬ[n/vQ6qD> O 9ͷ]s:۔ ۜ SYc Zđ|8_,*zrg7p|s=E!7庩rk" zӢ`X7-wh]\X’qd9q ~ۛ!\x><'K.ܤ0O&( $AyzS$_Ͽ~~/|ً)3?+hi 0\7B07Ϻ<jYC~dR+||>!+7!l +gW}wBHW5vRbPb_P]P}$`pVDtm ʠ,u.Dw1`>uP4xЉSTЎhKS7y`Msb'19ȹ5947XIr(I d)a @LLt6z^9&W dkWn-CLSʥe2A9sS[_r?$[D˸R76Q%"J>$h!FSN"x0R栯apvS {r' $Z/j$FeN ^ Zsn}7>{` Y+nh W$KDC},NJxsz⸀Sysܛŧ*씗vU^WF_=j5 KZ].&Vos(uS1ns& r[ pSD?݃Um`e .Yfc2Mժ֍U]y:q5E(\(%"rϙ\]PQ~NʽZ:%93xE (IVDPJgjVl:Yr_>_1nw*djK"吾LY`.D}>ޑ|>hZ Jy+ 3ln=iL9w)+#GEG}}`=AA˒$[Y`]::ZbI))p[Le13# ~1og_^Ŏ! q?Aֺe=|}/Wqv{m=cSpLBO.F]o6Nw?ņ<>)sfӔSS sr7۔J|TwlLwnAKzz8?2@!_{4m1 ښxQ>n eu佅w&{QOOWJ` p.[V4y/T\)9Ȓ8*T@Qr_սgYx_s觩afm2g񖉻>w5F8*ly'@7:Ěy<ʄ h8#۷[#Zxݫ/[dD+Pc|CN6Lɗ1\􎹴%PL¾e_UE 6bH26tάURbTNgyIEΰ@FwF݆q>R kxZUNbu58=F:'кÊo84 ֓nO ?BS)b@EmX)3 HQPvuZ#bB6ei Ի j>E.ĔuٯI%,BU>R&JM0U:PBA)G`[1&\D *3.( ܲ-Jt.L>6"Լ|MIwZݟ ҁϔ7k-+ւ8ʫN>!glBSW{Dߔ>+% xO&-gœMNɋо9KfO΃bmTbBИ` @uuɋ^ [ ?/q]v o5*{{1)|ܰ&Er.Y 3PphMš5&l kMR͓vUޫB3'@դ,[W y,7e6Ŷ-T Wu0xTZm-(q{&#|έ5Xc)!zUuΞ~6qz[ UfKGb䆼1jyl 5͋g@QlMU_m`C\ |{- {ΩVGBY|!%7952=#' j˓'M< j$!*d걆\; Z%jߋnK(h8qtefp4:K<&[̾EvEC9 ,M +)ʙBۅczb팈5lCE_Cuv> '䍰:fompJ_e}NxRq8< Rc ֒HcnPÿ$׋6f:٬x$ SRGMHUATlL&Κr=],׾g7TR($NZ' i`Rn;1yXD) KI{-H콪lk8㤬T7ٰ(_xRVh`p1ype퐟: bܐJ!yprw%Xm&5{2{J|=*;;/ke>ҪzaV9G{s)Gc08a*qCހS)e!E'"+_g#t4 OBvvW-gȂ]A :n٣흀4;Ҏb<≅yzTe}rPR nE?ݼKqBGτB>#_$ap2  98iA&rF 7ǀͫyvGP˦׈b}姣kigaK.w3W]]h@cyU< mEr.!چ"Eo>릾2ӣK9MWs%<:qN~ncӫ_>|zvu!5o=:u(ZF*t33TN$]:髮~ZԍER7[pjT:AgT3@*ˏr(CkmH\֟*g0z%8^CZ˅UkBYډ셁턴 iέֲ#"N7 f@6s hn- )?Y>3]*X>[r6x|&> %m?⣆kz_qpѮFinr)~1WVF־E,rg9pָwD>.2u$ gɱGe\9/XT=.~\8_¯I$5e "?HlPgC=ZJ|Ǜ#zmW:|.-"jl" e.mi{۝^ Q/Qanj1% t2h\$6 m(j*,2{N)r;?psw'Y?0mv0Da@&K.SIYqeQh`@pȟ} ]hCP sƵIĉټ* ūé49%[@T-b2wmH4ܥ\ nfn1 >meIG&b!ɶZ-r?)HVXU$t3 NV/  $oN׼&nd)/t>~_bW6ܭUgR?gas$NF>% 2:3/qns9,2OU~1|[h~|45wHc~HYSlXXG,Il`1IG\Wbb21# n1Eidmxn6F rmtp)ͻk/EƙOx-z!J̌T G暀 cAy"0q< E0a4="N!#0AH[ ?=*t2ݙD3Qɝqj*0`"{&CD3DKI DpC@,s$)v6YczO3 Wm<M]n2UJ%W"xW ,TsCz4B#q8)V 'Ț1J"ڂQCAO8Gz!UVb@G tOSdE) xQJtDV'u_b{xxN"ЌϽD<{M^y[ْ<=N}(Rap8F4$*`&25C4b$,va1YO{Ź'V${tJߕy1 Rq+IJea6ݦzndR'HO 5T\4n/yJp`P+tU*J%Kq҉yC :h욪3v 8]"K3eƙ>P>_'Ä֏FrQ\4%oϲ̞tVu6eJҲ_LGlC^ vP_j~,9`Lvn7Y28~. I˝Kf ՎIïK(ҫkbъ:]OkX\3T1IU8:RMUV8ʠ_OxJ6Qedv풆j#jѼӮc΄ND38D6T#W/nJpcd1!;mTchyϔQP:8by*ѷъzOgq6igqZڐjaS~ge`HڥSPZXAY#DωZ8 H /%t2d/8[MX=֮^Xu3Jn;;tYr忺HyQ+g'3ʭ6IO8)r'e7'>N" EnVt15'ð=mR[_ ݎa_ޛc2(PV6 ]Yd!gz=?'gۓW%UxP\OIQm0»EZ{lx.y:,Yx7 @ T _g/o E#++?Z}'~Q!3ɰ7* cVz*&CzEY,sn֩iܘ')a30ԁj WJ!p"> JKt0pUNjpTpaĞt{[9W\z0U}DGzp`gYD )[UwN[˂V۠Gn\S޺ZK`VSm 9S:+V$nSwK~nDN`)QT!iDӑ(vNzŢ[YUD =_5ΛSI;ͮcd%9" cTh*LhFVˈcF ېUުnu+Pz<0!u)3`pQJ8N#FI*ux05*q8;\ܾ")9IE1cEwZI%HD0 AДHK&V[{Ac`t4Eg)bi8 kxl$ 8e(uT3d<*6p#.2~dgaO4#HAz50EG j*9icJBR+'Xb (u>VQBWÛNrF m&D J/A @6jc"ZF㠮Nԑ:hgݐT#b8u3;t}j!5#rY"_aIbKR$ (?ߟ%(T~_eX)._l` F FMWU_Dgg[Wգ[)1gisԭU 5zbe&Ϣk֊+XBbJAyvOU,܌g[Czv^9{H a\|bsyU-ճs~$ ]2vC C%XcK/n馩Lga3ч^(4i%@fmvn:e͝#hyu'Zm+%`Y:J[0KXbb<)^ȢJ1Xq8Eg?O:#noPē+~%孳3_$1o\UCt-tIWpe;u:+߾>}?^gj{Hr_%_^, Cp]yAβdk}f4zF55[n<,8鷿oVuaec.n gK/^oui!z!-#oyuuCN׭xB@1U՘}#\+#]GEۉzΉX=eEeIrEH*)7P8T,beȑg#+2-uXU eo_kJT s[lIAp؂75>)fsN?P}/͂ tZ<8 lv KVm흝7nUs0YE 3 ҖA5᠚B2WkTʥ{A`LHZؔ4+ooZjq}siQ( J )c\!E$URdb%R)X!lJмoP18] L1){Os8Cw)v2&˾o2A٤)b\S/֞f(֧LXIhlz%"Zу7b.1ŴFΖ;rWEߞv4GHgjjVbQ*dJ|gXucc$hԬ.cbB ޲QwA˹j9J@kR ?Ά@;۽M,y]ٖ %;L1Jfo8eYu^-#4SQZGy6&[sG?ύ[Ka}(9%@MJ1g.Ѧj Z`I3O˷{QZHu6"6Ħ.":+%.ǽHSA:J`\DZ[{t_"eX GQg-[ Dp8Ua)^dP+]qmQzVXo/dWo<`[=$aȜؖ$y66UJ.Af΀~GP@cmw:ic# =A?zʴ8墝7/=Kn0=Qn`v{}\w`P*JLA0J4a*vHpQy99tRtӯQ|(bҚ'"YMGLBAgp&;L u.›mDv:o/8'P, VTXN5U,zbŹ"Z2"LM+e ,a`LJtXBI`a|=wvXJ.KߟKotM=8SP%T,)ϰq}`46-+mٞhv{jJKy4`*6rv  H\kL- O'b^S32:9JAvm)U wgB~|ꥬR(Jb|}*s>cT"S^*LNBԻk[7O%+Q\hw6~,IVs&\Q DΞI|\h, iB[M ˩9Wvq2kDY!`P3LAMKqՀޯ;~ȷhK_ Xo \l*|]v[%^O0 흽i{z^_s[J~ۻ~Nfcݫ_=|{xtv&e/F@\=v GSI'N b](ԕE݃,ԝK RMyHvNpq#<{OS,SՅgm[ⷞ/_3ʚQֽ(nA U?88~fb;#'BZه0i"BzA9>~qMB*f SAa+bڹfC,aPf[%F(\cPJ0"Ӣ+d,WC6}y~ll$i! 6IYl^舲q9e &[ꡧkz67h.glgm?{!p*lB4asvrҬE2;y;? 7*ѐ)Y|1)QNLjZ=VEOWC Z)D&(r\|pI+ap 7 lRU7b`9L+υ7{*I't=N,9_6}ޞBW!jt)]9%clZj6&㉍&JmJ#F! eXJA+m%V,B͒@u mŚz!V20q[q7`GGس=N`wY>,BNàB _| onE=@iqP'jvqgX3e%_鵖k WD)fr b>ɪ7Gso܌`M5+LMʦ+ VF5mIF}_t ;;H~ U`jEQ([R`R%"8_5UuhMKNuE u0dݱUWܒ(ɋ7TQAzn(ʖH͉@$wkwc) 6W4T?V.tB]ƪ T-gmMP-@#C 8Ȃ(5Fl3RnIh&]߻1&P p ً@\s>YZ4q`—ho14NkWf!{{}ӽ{{[V ÁOL h@_O~.> FH 9Ӥr쁈Ēm8+kUE}5U]RvݔnxolKmڟsy% MKZY7GJ:^:6ρؼz|Hh4l J<+u"ux%"e<|ML"=ݓ\?3`Z1mv0ODal/(Lv[-c c17>va_x/ _' (;?./GjloUCn$n1+>_Iv]Aͤ W;کBdJ֯G#^b& 09pU?kM J}BOן\/c f -|ùW{,?ZsoΏz:~qX ˟g^/esoF~W/SmtBT/6x)۲ yx0Ms2^hImSW&ԧcrm΁ﱢ'2~.&HGɲ-0,Ԭ2*G&4$T+19s0m/[# !˭|֘^V\X/laX"]685e8 wfW°w:|(&zk$QMZd b0dĂ[jΣkU6ָ c3%MjRj^ ]R#Glo MGA⯏̕E>{3N|SS[ p!BhTm,3/d0֏DOY DEXuzTª4cl "sceA-ўq8DUa+f_U_x!Q]E=f'Nyy„Gg'߸vJP.48J(4![ ز |Dfp3P 1t'ruS)P ], qP&fl8{l::+sW0x.v`xW[t@l%{ַ:&P5%ClMQX.CĔI)CVdh[V_ukGH,MUR2n1Ll}n1G1#=q˓OV\V rYLE`%lzhs~6#$*ZTX$e9'SUBZRPکS`نꨎ&撿iItn\%v__]or-*A*.CjFZ,ћ`> |Щg[ Xs+h:[~q~qW0x?t{?܂ N>w5h}.7'k?Q~>x8]I.eq0)8z2E$7`6^ۘεUV>g0>ɳjw@bm"1""@4uux&Pܶ}m.|~dv>i ??D꼻[hv/[O}6rbAmA$sXBJΐ*aQRQ"guffCٻ6$WY}?L 2/cu_62%&ElQc%dvvuݪSnW 3/"|UmIkjZ5L7L[ H%=$𝑢<}%|~UpGg< {D8:.pM-.Fgq1_#,#CdAE+w5oM32&0_rR5N0-cz<:$Tʥ1+4MѪ[ Y694dvByl]`s֫bgo? Lv u'R4ufwv Ywr"4'}>̗z{OuiU/Qr Z*.I#diYj X])8 w}JӂLS-j&,vG o NP)YH}iZW2BV7MwcCKZuBYqwÁGgEg3*䲘X/uAJMRW` Y؟3xs*Inr\ l;_iߘ~Y}BX͞R-me|!os#54G0.ZWP:ta΃Mƴll)?A:b.4?>ٽ61~w,yxUSnЍo<|p~TaL'/;,cUc׃dU1yї+A~K\|>=_YW:W~:r' lMryԙ6N9G*aGbpVnH`טCLpu_e͗{ NfMc*fOыX¯V~o~Gj_FEAM3L/E/M/9@Ko·bH~ae+nhXF8{8m1sܝ b~V:E^]]6u`8yq^=7\h ;\ζ3"- 1ŪV5ezc<,%|F)'slj%*&L:6*E}4E;z>CCܿ Cay؝[$j-f QF4}f:mZ*rIv^\%(:uchu{׷uvW5Bobv8DR;l6m(>|z~nyq~1X_Vo_Ǥ{_ևg"8"&_9> \sƵIo0TFݵ边-"#JTF-z''glm9A/tnxK&G7W!!扒a*,>s{sz;et"3&ID,tYJ%oM.jCz KRim)Hfm،t4fFBn*%w^6d^B-ߚ-xÌ_12idi'Zm+Qk3TifeO2?/P%Cח O_O;ZtޠC|}q'-nKM흎?S2X~8L93PF\ŕVQp8~ାvKPiӪ߻:Xmgm `l{m/$Rެ{4~VO[j+|e>#ɜ[UMq)VLZ+Z.BVK8 < $] jOC9bavkҮ|-qmH燋R?5T#M -ymHV-T$6F%,6tuP*Q22T-JL^BւR**q-4/0w2)[ZqVzj[5 M)=+,TRւYҩdLFTp%&u+Z)|RnTZ6!wcԚ$oVH)) LVo6U#|at"g̘fpfЍNYV&,j-'"|?;,Z˜M4ZujeY< ZsS!S^5fBaMDnBJ0U@Ͽ-#=NbUVSi-cd=;_PGBF@! "=A\N(׫,ƝX+O@ɘȺU>(x*K/jjƩBjJ%~)A4r-DJ߷J0%(mv)퐤 m@qu~OеR|MIg>).eBj%6Bb6 R^* (Q3*,I:'Y笃S:4J4N6jڗi f;n訌 +@]I5+l !D!: X;UR]J q;aJGL8 F3U6a5. [݊NVp NkO8v 5p;8V( !j`yU+J\,dp/*OZf@ [MhlQI`; +GPfa)-%2, / ѻP-`P|-+MP#R= kqCc]5oO6C!Q1`.#Mpe :B$.qJaV$2(J@Bl:i*XW. (_ ȱ,Y! +u zZFk,`BܱN[ pI ]z5a #MABXF @$rJa*,|8&u0 #B c&j|Α2)ؙ5 Y f!2ؤΎZ@*roSBF u&+%Bn7*,7j{aH@:ɿ +m%bE.5=))E̾8ppZBTCU5ЯIAQha8HRX-@E@I v/UTz2V"P% ywA>NAڵRyŬ8c$@T6zD"Lh {`V9[+WmtVU%}7u.fM|ou]$h>Hp"P ɇ+WU=P7ҧ PdUI :A2iZ.UF6<܂!&qaGe&)t]8gԒTx 2LĴbyU!}ڄL0tq\X1GT$/ɓ"P Iv3+76Dn+>IXdu~R& :UsP5VwVT,CT+x-F0Fxv)=hAjz1i$!9:]%_>0%lC^J7S ,C)9i 6GOրh'If89.Az-l l/ lwLXä-D;)g`1ctJߒyAP΀Ae#Kr0qkڏ h`:L!#EP\ʀV;8 rZAXd*LJBH HAV\Y4dK'r,dxV 7+ytRB9 h% 21"RLQ{)" j ߤn%,BG jd_lAK@0lrhЖ\cs^y^mC<}y{L<<^5:Lj8 C 649 4$tllafݴ{GQ06,,fwm+JYV Q 9IF . dL D98"hl|6l|IiFlbp)^"V 3!*SA4G> Fns$IQ.Spwz ޵q$Bewqv_$'lX džзQBRm`^$qHJKU6HfwuOuݺj(X@PXʥӃ!6_rխΰX R! e/V+ "&ꬍ%''&\Rm$NF1  Ʋ[ǰ@$> ~TvFlV1d[:A%~aVZ(K>OvD`MGAp.]+X 42?e/*OW8pw"A"xLlS T/rqh4]B1G P-c*i\X u .SWz)`9+΁CUnv ံ-` Н>`=Be7cv r_@Z+v TJ" @TyH D!@B$" H D!@B$" H D!@B$" H D!@B$" H D!@B$" H D!@B$EpOH XA ջ* G$KDq0} " H D!@B$" H D!@B$" H D!@B$" H D!@B$" H D!@B$" H D!@/ $A|H +DD$3#@B$" H D!@B$" H D!@B$" H D!@B$" H D!@B$" H D!@B$" r@ ^ $ x r$Pk#%zH p7E$" H D!@B$" H D!@B$" H D!@B$" H D!@B$" H D!@B$" H D! K >^T<*>-'7 wCJqRr8}D&۫+`!J7%V>p K/t0>La{#*rg ~/2z*+ X 7*F\ejuqTjFP\}/⊓^=? B~>iN3p>MENa=/ꕟ^| ="$?"?$?{Ѵna5rȂ҂GTIu%Lecp[YRo' 'A:l~V}TkV(ԧ{w&vPtlrBH&_!o;hq51ٚ7QCS6WL":DEY@l᥷Eَ`^'pEu8#m tom&W}ѶZMv]f* Emm>7F\erqfURW/P\q'qOW_ꝏed*BqV/S~,?ONS]$%<~~uxƇ!,հX~0_UBs}b|BA"jZ8B!(E` ?}o;d0<.y;]ZD,ONy,J(C$oXLDJ*C/ 4%!X?^\2QZH"ëq3m??=_.,QYp3g鍳_ 66 DcnvsxwEe!zZ%˧:FW{n&簉%ޙ+|0ݪy:-g/qjS. b\6&[P/#dEm >(}7ə C7Mqs#mt /8W}P@ZWϭa9 ۴wUy"}~5e-[vyxsMP`Zh>f-xLc{6o8:ʆE5p UcqS?#}fӯ6l;_x"66'oI 4LBr!̔JvrZɤ8nccuPe(S·[-kŚ4}_6:~:wJne#q9  ^+aAb] rr?|s>`A'01Sخ3F`%QJvJl 6AbFF+gEr,汰,#0\L)Ʈ]6}d\ol| -5<4W UEhPZ:K'ܕD`T':Wqe>:3a3r 6 /;p`/nCi |AhqKO`BWpas9M]w v$t(餤e"uTM97j6VwX)>lK6ՆW-`=is5vgj-j{P$Y۹zᭊu~/Su!LA9#[ڂ~LrSFD7M|+x'uR 6wpVa'.]p'z|j<.g\YMf7:]R#A^Μөm-村=Ji륈o P8Sowz5g-Fma"YN`KO U]u8Wj/)QD6羐h +>YNc͚ i4KR:$ˤ 3x/C3#)%,J޺K)X(DRMl|F0Ds`_J9/ {m$%W!䄥7R?t3~iC|XM4&_I2iEn#2_jSN!$L9-sv^IaRV nmlz#?7~> 3؁*Оbja4퍮n99)( r)/<$Aآy6Q2XO] $yRo52;'7~釓I4֘Bq:\LeYһp iV V췦JN/?*8JV`Yw(5KBdaJU#ލ95{uf_po.]bzsXNvK"`oY?|qiZ2"u$!Wt6 ka-lĘJF U̚|4|-d1f?]79>9F֏:dۨms9u:lԷ/|4WP ȯ#FYȪN)Yȱĩco$/ǣAjb9%`)0+=X|^134]ܕ5UN}_O?B6~1~yɛߞPfNu|7 eVIQ0 ǿm?t^bho647b;uیkr˸?Wh7J!fŅ+$ru!nH7~R:QbX]pGWT2xi_w)ꠚ-Oy 2KVr[S‹+I#8IL5>nX'<`gd߅<.O6C \~b[>ڹފ1;;lN[nM:50yryʗrj/ANS*e A9sW[Oʈ؛Y[[׃YJz@¥k+`$4y {c/I |I5l_+.EfB򑔌d6wᰙMj,7y:ɋ[Vl֬oѿy&JdKu}\1j>\h`B\F SU| nfKYf,IeKRjUrAKW]Z:IN E[:f#5&㠍9`5 hq>pJFRTP쌜͆b#z|s+x h̥3uQ|6jZi3٫koPHsv )?MXLBuBYd\ᕐ09ehb@yܔRQA$#VZi(U΋hZF΍GOIx:5r9t$6h}҂O[1 BiTSsMT%ަ*UY'봜_R):JMx}^>]0ovqV3YG}b} :ͶI/R %/]I6j R%K(1 '˄h$=Hb>6t"TbeQ6|;my#r*r.6ߝu Vݧ`~]ao zSy6,r%໷m~7g%2Ӿ (>7¶ÃYev}Q9C-ŨZR7̩Mg247DfEHƔo1r^'P}|0>19!. NZ~Pg{`19'545:Be#)Jf8=D/W5~}ޭ_G>{OG{F|W_D۳٠F?N (dzۏ×WU؄RkSB=p֑ŤAnc2o4VFp6 TT+X`klP E)a9;[B#x9G<l@{~VfAS bV,.6ペf)yvSCIiX,屚owNw]~zvMJ%_U1'dP%7 01KDѥz$7HfWZM5Wmds)&ڦM ):*M$Gs;<p}:[TbIURuZ4} &hpr+#tv>d݉u7IEٛ!cbj$=sʲ%"; :B۪xzFhjl z tX4dk(EA&S\ H$kgZu\jҗc cP ً@rn>^ӔD=%:ɦ]11C1vħM׏k/z:N=Ma[M i:u:|:PÌzѰ$ZӪ>jL(SkAlJeG;>ї牾yooJ[0 *2-"ER+Bb 'Fu 8v%,hBL\SˎCU21){T׶+̹v %f7/{/{7.|j?$5ɈS,V_N*]BxSQLtfma;>a(C+%zf%xkmoY}z4= KX0%l4]us_H9%"90sgכܯ wf˖J@&o(e֬>شm$rC`d|~.;cϩ3oWI&>>7)9&C3;(M3Ff'6J!j4gCP;'6E謈"K ɶ=m_:b1.18||鞰u~O\fv >~ZdP?mE |;0m}%GAgt뻶֔bPLPPj+nt"˞QNxMWA/w%ȶ]&27zh,'IM K@"BS㜌M]amtё^$ɲ _oLK% q>ݔ'g s#%KygM'0d d-)`¾NV .*!o8 SQԡݮliˋHEqTf6YebIL"(b j89`|(ڼf>s4Ӎw'H_^pOJx)ZTJjDdp)]Sm8H웮[C0-o`T1ZcRB V| T,Ė 2̜ ]҃^\J:cUřV{MKE2JsW>|e|R j!J`񪼉AtW\jrCl t@L4iۭc}ΥMw=qvV֮V.~h=ByS->$Yظ@եICuk&xdRqamPh_$Oܝw܃>V%s>EDhg H8\œ=—|eFZN%bB;Fl¡ NlL:͛f.Fyז{(-ޱ!mn~ÁSȱn>6S*&ZX)kLPMZ//M[4>iWNiȋ^Q?h=,>H~1x5bqzWا:mџ>:pp`uc"lt5#NR:鿿$ fӥcf]lltWF#LE[9/QK9&:.(wnGԟƮn V2fWoȟ`'bO2y=Gwr4ǟ~F֞Y.ԗcexMsgq9d[WoNݫ_HX4O.Bח^f=s#,:6zx{zhV߬ܳ8{Pgg9>9:Xo;9k։:_D~BqOGkO&;?Ws4!%e`]Sy6E$7yiEVҾm!V[' * yWPt9 zM \^ęgEkV<#-=9?v[wc7ή/X?~:>fkh/(n|Kڣ|?1IWlW>j) rZ$"lm4_:# D込{g`x9x9FSݜLX2WN)-ǐKؠAPj1l Q,,9Skh+].Epwh3w SW70 3HKPw5*^٩?KeZoafc^qhlp×:|6.9R2z+y9З HA EI&I=+ /6 tY?{WƑ>q: N0Badx->H!)RV)Eu^]W}]ܢXȤ* io<]֕p<|-GK'ҥt n;D)$`%%-,`*?.SVi?ޙGѼEW:@{ޚvOjrMġ0Qips x~ܶו(2>|t\d\Zߍʯyqd6'fOT$^[~r{CћTN`8漲hwk dɫAު-<|jmѷIݷI7w>|A'ۛلkɼc,~[R>5x12=B+t">CW~>.z ۢ`M~qq'ZM: !Wl[Xcrkܧr# qєJ;ueM+wǛN' WL#3>{7 7i=ӊz+LŠ+f َEhP7 Թz}?^4i;޿mp;nX&bkI55fAmrLk̸&ߙP\QPڛ,ҝJkh+\]%i9?tu\ ꑫ+a#Ds0`7){P.qjV!YcN- Oﯞ׽КC^bR"Q zL1GF)W[ jd:WEp/ƦlÙ~!Ӿ?oc>~؊^5x5hE=s' :,OF % Ƙcd4dB[NH+"Ra <0%vBrCe-&"a "X$Zcm{Ѷ ]- QG}]~YFsGN?N J"7Hpy V3cɜ# u0U L'toөjܒ랰LαzpSHY2}ie f5XmN8ugVBq+mԒedp4Q:0Y0)tx7;:4 KJf ,ˍ8?y4ZH:o!uڑ \DiT0o6b$DZsC cεe2wحnHFǤ"}9<1âҌ;$fD00颺LG!]c5e臚0#n(UJю ej&Ҡ G%&90"ye/.Z6HAz5 ="TS–Z8Ek0@Sy"u;jQ0n&D J/ 6jc"ZgacN8= |0ty\P W9+\ݸ\i[ D !Oä|(F!QMR|6%Ja<0fо\"?-q+6(b @QSr`}1,; @ڔA#L9S)(8\ˣQ8e)85l7RwD=_BI;?5 Yڡz)\CS)e_5Ux~ۨyTRVkY::r"oVa³ŸSX+` kq +ujRuVToT/~NΚ弝/̗^?N궥z޶ $0L":3Y㞇KВbKKb.CCbˇ!J>_ =ٻMVU.AKm+eu)>+uC;GRσKYX L1﹙Ʃc}Soҹe N[s]R߶ is}ۯ4+ ]MZK銪 Uׅ `wN^N7'?p:y㫓o~o/`:Z*K!/MQWECymTMMoRKzLyW*D !8fr)uaZwId$쌣z(E$WȐbEhJ8 q^In87ZRV!m:HP>\bv6&5KYft/4KyU:bVeםP.zNyqߣ\yBlK*ĔOAs'+{8㸲ʱtXPK PyM\hBA:OEshn\A'!snlgڝ[}/){L>}wJ`8Xs1F]sSO ֣HPLzĽhjSm@-=N0+rܾf_&Š&xIj̓ڴL G #z T,ۍoB- _xVY} %y=]1y$djRgwE{WgO&q~T}e5Yԙ7o~=mmGoa3ZCūozI>EW@DxAGr0э HHc'z1ӻI53輛;UH"%Ɩ [q 72-;UN!sj+Yĝ*0:jR }i#r4Z{4rlMO< [wwf4;COJSeAi)i5f4F†ZOe$X*ƨrRgGMBBX1J9Qu3q:A-Ü,Z'I RxNdL%\8Ц0J#/ .of^ɲ2 CoIsخW:U3 W1}[xrAllz xYE'ۣVg9f؞{:G {Jl]$v?G`RJbxi2C키dG(_G) FQ`6  {hX˼"DKHEq/Yt0l`cJr4"Y@SKknXe;6j:g4߆:ʄk8%M \Xd< |~vqў#b;|Ϝa c`1X9 ϻ0inF1םDTĞ>}| >5ڒeϯE}(Q(VHdr&+, 4*2O ,^#6RvWPkܨjZ_N9Vd=(,\-ndMZνmM凣Ϗeq#SI^z,Fyq,r]/sYXky+}irKI珞.]}' \mE*t3Wt:w&j<[+vƠ,Y] 3Z4zON=9Β@YI%~\ f`&:$j0CC;jN87AtԣtWӕUkVYzVe=mvz ''g]w'YiHi7_i(6 -S夕o\~!l!Zs,x0!pD}g;HSs{uB=@-ZeT~,?G/?c+"LlUTL01( 3 VvVP&U{墴?vk׉5o{ug]'=¥<٪tu>].߮Y!1U 2+Υd,95xҐe.:ebV([S`zpbmkQZ;C1)5Ok &gK<NO/sw7MֿwHOGbUDD\1G,QYwgb0i5f .7kh5no뛡`~E;9`3_%Aϥ9A"iV*@u뜙--ن~2 gJq5$$󔢉1JQ䪏X͹PB߸WUJD.qlĆ8"딵IgTm(t︷n^_NOX ;$(؂'.Bq@l4Y{ 2{x՜!'`zVo X|]ooa%x \mJWJ8R9éVȣq#Ɔt寱xG7 ;\gpqyXS89/ dX]2/-9ݔh' sɁMYyr=Ćx:w>'ß:dԒNu1eSa=^zHP ±e%%vDg`: Qa1>֘,U& U(0_)/vζkRrDm1|*gM:ѵ̈́j-3,8J'N׬"X[Z(r?C U QUU@ ֋D !IXkAډTNVogOH:$Qhw^pK?N,-g-,c NɎ]U#8ݹU; AP+6wX.ʛqӟԢk~ym*BV蚝EhZBm+3:O5mK\A$˕#P3Akⱐz3c78[Uqa7  ʅ7ڽ]|mx՛s/5uFW glQX+dR,Dh46FƊƆJ׷߄Z2(#hi%a&LUb;rAv5&zy2 1ݬcWά3k IGc:k-D$5P`˜-N:NeaB^\J"dQ%NpM.*9(JDBf$.zo9w<VH8Fa^#ґR=fFqS3NWp21#J:8f@M:%fvhű T䵾^+ՎohWċ\ɥLL1I@kb|J&Hcz yFXdv1uv]yQwE=̋OQb$@b Ų5 #'=iVlA[$μ^Mbb/ؕ3n| H_j1dFq $9@ُNԙWج+N\3!1]X}֔mm͜G:zcܪxǻ:ng'<\;-+-} Cv=`N:6]V* t7GrP"o\_iw[/S (>F.h%32DIb I 5,er9r >Ǚd誡+MW٦sy :tp8rR_exf3}I-)bMtH\`v\P̡~>÷7̙N1 hT,{0`%VN^mEVם.~p%`_^Xm++qj6jNqs{5pzoo7v@~;a טljC㙍m/찫 \iQrA5֪x((cu(E9L])HPP ,3UW-;zܰ+\m3敝zܙ}4"OjmB+mMI_ %e2eS53W򬓧XxKRqzL̆* 6zS IA)]r͂(J9}պ|>{lw:j&Vgg074\KPуJ\#in-܉JdUv Kv*0Y,g%~+ˋӓyyCmhBLR\b&(R(ٻ6r\*zﭺOgw co9c+d/(KH"qVZT7px)ek@tX/BES✝/ak{kC1h:+2:+ :؞o{d= Y֡,P,Ͳ́2ULW`nP䕫:{l5<8aY;LHNKbBŸ uJ0;"džܝJ<僞{7;9|zbV9 M{DYVډRA05{m"B PJF?de1t 'FZ~#{l8,L#B)!u&5݆s|w Hs8Cs+ ''ܭv_'{$^ڭ^Zmw;9N8W3;:a$uމ$=$_;2;(O^,g_km}gKe//d8}?u[^(N|fcxѾ˃p+ ril<S/,fK9M\Vf9yǾGl] >nkͷ޿ j+duR[:up*(+Q*#WTsl:")rQG4w]Z{M,kbYOf7G+1_1ҵ=NL땘x!̪=24?nozP*W1 :BBLdv.Nev8ǵ;B'ftM])dCqdYU \53:# .m@#ab*sԖTb૭-3k "b(?|PK݆Ҧ ǴN_lk5g}Z%1GQSi~O>&J9m~J\n$G]6T.8qhHEàkH2_~~~RBBOPtT 50L.!UJͪc´+V|jK5 ,j8Dc@V> ̤bgm8G﹢W~}'SfW7A|W_̵Mޮ;0.,Yy?~»dO#JTXY>g 8 ,s1+eX ]=8GpDcL}[ ST*f$hU\FP\Rg6=x'z~|uy{kpX=+Ey@WLe[Դ )jچ[)j*u^ԴR۩/XTf&]}0/xNuBrޛA,꜅Aux:g~uDfDa"39ۭ{&OrDadr"_J ss }$Y!xW9YPY#dl!F r)AI wѥz^yZkϗ<ւ4$`kp#&bNYqJ(LϚSS)b6)qk {S9TmU !MzSlʇoY0l tC]6+sQ (@h1(.-bV1ԔjԅSu !QtO:иFy_c<952q!`Smըg :]js \&Mu)·^MmjA+^`͸w |vխ`ZGMꐳ}֡ԩ)7N(&lVԷUVWHBC+^}FLUX|2FՙX7EPrˋWX) RT*jtQ9+cP]Ƚ\ i0tL.tav2\Բw+{g<3TyBmR!AyUhUB~ht!ŚRA"qZEU *+%nbYWVaRzOŨMg?6#lW_'e_~8drx4+\ыe*& b[ؼx-*[S \L&7@}NvLP&F_UθֻF-Q5␜ ]?QnT+/NtTS!Yb\@km`Ym5@[2*m'eXZuz~3Mq&L2Fud:,p$̕s 6fԽ~툈U7<8_79\g,p"R7gXj "2TSV$PG'cCu,BN;r&֜gt}!9eX2OVD?$$L&l >/e.~ r=Ćbۛw=2Qjr C(Eʬ& !s.Kb٦:K͉cyK!kI="gMT);wf8@+V!;|L0I/C)m+kH& DZEX^U1T0!UJѤȷńlpl>٣ ;Jkvʰ<'FjuRDjTLU5*YE{-uB"KoQB*z/Pc.߸$!K$Ag-xr,x0n?k8GWBKUi`<ⴛ(NYIqzr(׊&{Wjzcڻ/:2T>"U.NR*{g=&rЖj\\`NsWd*s%Rؖ=2,2k$ͩ0jS(xνp؄ߍtn1, _ GԾKc{*=jzFw4_/5hEd&m jYE9L@  فָR@܆&ke 6[򖃎h;jRUػDB|U .6$N8^q*xO{}^O9;_CViw BN_)@Ke?nͺĦ5$bY! 0?fv3SK=&έZlzR:͂wbv{<}YH~{+!'<||Qڰ~ MbK9ȁ͙'y\ aw=Vr}r;DTkԤ]A]I ̞)Îw#7['>dc!B;IW0*KC)r^6Zfn^Z+Y[uq} ]R688TZLS(8)][˩ {:>OAO"l\:nllO 3Gxgb`}q!c0aW^=~8/o)yTfm|,fK9M2'pYœ[n 住Öݳruc͌ix77?'=xˈYgx}ޤw=ݷ3D^Y?ջY [-ϕY=#I1i Cc<8ce %8m6 K{п$A,F#UOYn%6 pQ2j!#IR$bqWGW҈jʞn ړ>Zrsa6ř0ޫM;O}3z~b[1ilb:FPZbAgWz/NYHq- :!BI18GF~0X|1K>=8pl)'+Ԣ+0@he3 *֪?"`4ڻ5 L=̳pX$1!@{l!Z|>?v=r4o?"9:?Z>a{`TBn-\lJ"DQjqx58[-D"$p-D x'z&yocהW1kSp6ٱQ 8Z938q AMy]Oyã% WF7h=[d @N$1DރR+AJ ONYR=>{,AzjP?Dz+ zzrb?0N|7^eX]6KP^X3a+bPu! f(Zau{8szwFjWvlv\y75Oi[Y ݗiiVuٞ[nh #I+88޳qI.,wOgTH)U3CzPc$΄ iv=uUuU5L!tQR_rn? Xe߫ϫhZ\T);xw#MT^U蒗 a:.ONG7$ Q7|zJTEn|f ?r>_T\Gy dʠ)xXȿ*HF} rE5swZFo._HM1rc[r\k??[E_s[Ժg:t-yvr5`o.G0Mk8jMQڙt~V^j%4C,͛>?ā}[?8Y_ySk~}q71~]XNrI0+p̗|}qtx۩d_UJw}n29Ox_:M+R0+ox<:[Il/I,/axhS=/f݇Gh޸ jΥƳ^z:o ܭZ7Rl:g jJ=<ճ oS#1z:!jѧ= ͵N%Aa4i779hU녆x'W^$^!ΘfMEҷ, @^mvH",J ƖVl.9mcA-pku 2JꀙeR\$i>gVY<|o3Tᠰr..ϻ#Wõ+vz󆢄ҷICPg'W3s\uuź, kmu9rTL V`ƕ[ .Yic%2MfSULՃy15SøTQLP4KD3o2ƻ`,fb GebNɥdfoM)zAyoLb:S#u;˝^iTbjkpL/\yQ;7IiﵭDIfمׯ }@ IGtCAa/̏fv0=QXKC^Xʷy+zw\}VWۈ׼T0Loᘓgq~x_[MT v"ȁdK~+Ssy+ s޺A4+~􃒾+mh*&F+gerxȤy,-#c]T_._}K,i fm<  JKuV0=.aJ*y-4ָ|p9c\С?J,ͪ׉Dt/1Px(S'hIB(@`Zwh巬&N# ٷtX3]Dݯ ~tWNuuqγ.~-kOޝTS䠧u)LA[ڂzyrFw ΢ʵr6?ID,Viu@y%@*kgLR~E45[144Ⴥsc۞G]9VQZ%9I-h< 'Bk4VDrW#j ;nXM(LTGt/ֵ.:&IY ,&xcxV(xv$٨CBr{0ښYHIGb-LD#"9 xgC^`;bra!oeeiINALȭsP}4gdB$;]rU|5 ,, =v95 ?Xfnx.UjiI.9?;FlJadXF9<]8 Т2c8C Y5 Q0J('׊sÑҙ?|4<;Jl5z!ɻ(Rw7\f]ᅌIe˧EuUo绉 \1NW>ŹyKf_k꫘*ڨ0T <: k&?B]_/a[HHτ~W\SSy ~T?O^W|rY8 4.H`D"/4_BVXْwtV6(P(Թ޹AOzjÜ\j|alɸ`5$`[a`J]R ')Ip%Pzu9TjĖh75vS.;4iW4!V ;qݡQhɦ Ww~Z6o'e)x4DRpPz\@S2!'OvZ{jE>]<VI&)i,PI ||$axϰv/WMo*Ԋ"O;qdpkRln|;7{o:(<L&<%ҕњ>V}+Dki P6qX]]Ym]`#S;e]!ZADCWb͡Gq,y_,{ks=i\6A%Ǎic0Qy,1StrQ6OapȒ؂cI:R_W޾ߠQ,I=:B5݋G?p n < R{T Jwu4իGe=WJ#9izl1%el3B @~k)f,J1d,bQztc%$n v[JN-5-F;2a =+mo >E ]r3]IK]!`{CWג}t QҁTfDo ]+D)@WHWRZ%tRu ;fhy+D)!ҕRLt fM +DɆC+ տٻ8WRFfG0la}^l(vdc_4);"TX5] /3#_iMtCW ZJцӕv#Bk2,WCW צЕӕ7f+Қ&2+kWd\JQ.mg>26.F)ө7 Yb'Pn}aUyBhRU+K?%=qF`ݫL]%~\B>@yޢp|wu}q`0'!ǿvwl{/g7ߕ?zvy99+.tsoq ꋿ0wVgv}]|'oS4'BKK?ӛ>~KȪwߞ\P+ʇ+毳3e.3p@ϋ|umy}fG g43;>Gz}Yk}} Z~}Wf?^և"=[N#it ɵBgJGHBJ)7:Oo!YF߯^>Bru|o~C1P.`GjMQ"ps4EwP{OD%ԝiR" )WLnFzc.ڜ6M >Ov*u('~^hPBmc! ͭ4{Qg) y̙Q!v-K"c9sPTI Jh !Nh(e%b$Z=ˉ8>4ŻwCRMV qxCd"7;|d@IY|'ݘ!>rb1-xtf46suJcEĕIF-9[ w@x$ ZYDEb(GtQ!lCt-2T`"?!Jhh*r#Qe̺敃oMC#%C% A; ԠmETm {J2@p:? v[|9:"RX{bJ!CMHpm BHI DpP\ZB2 W4-~*3\ +n3$Xr 2yҠ yB߁#7CAw^KC 2Pư)0m2 19:  AdBEA2}RAO[[Kut,,x$PM;&I10rXhB`@yNBp fǭl 39_ kPA*l*xc(@Hq`j gYݾ(EzeH_uBr! 8F&;g{ Ϊ J*HQ'B BFAsy"F5ΐ_.Ct@uR"z,s͐ȲA ۣAbS9>TVcF= &O t)`륎_ݢł4͈ $'ͧ1*KCw"' }4`w7vu.8ӍSÅVw}`D6# X = bPT8xiLd/PTdQɜpt56Rv U d`V`yR@ q0_<+XX肸az+R|$Q*L&rZcd^[`T,L'`90/) JWldMKPj#VH܁m c[*YȩՏoM?vAyƮ;pl2|Y 1Hk1xwӿ_^jw yv}LRj++tmHi,#z,%\O\w.x W38Ž5ZSX4kJ!Վ a<DG7HT}@NemIg$+(yFh $2Б1v"Xά$12FJK!6"N7 7KiVRa},8+lA ۉ*qPkV}Ot,5ip,jVMڀJ5[3W."1 RoV! qB}Aosɑf nfoyMms(`#;ٮ ~wǴWgu&b́ }DH7?u9AgjM5g{h@(w6xY:J;5WZsfLHY#Gh4vfcgrzQӭM Jc8n0Ȁ]€pe) ~Bg A-<&:Y;%\ݣLv LрldiҳJG0p}f=5^8TVho"X$@:e6ᓻa1wHX j O(U[T1 >J͈T1cRuN` TA౮1hG1=\E->Xf=d(P5̤Gl!&v8g|AB*!?]ɑk68Bw *6lCY=C:iQ` EY;Xk&; Zf`-|k䋑as!~Z)AH@.#45`=]GCX4:%قn-M; …02 6|5fs˥:pU.M˩XL:IoKC\.X,7wR@ŵ[ YNt!B`︥EB Ƹ:􏍷St"ߏsqTws\ ,츠ѬJ sOOP]kwY_~y'otn>nq7fE/ANOn'i|~3ۋ?՗}Agן_ݵDou=NOwz{}k?:zi̕>˳'?{#Oil~'ʸ/oMܤxj?"ɮh%#WHl hynQn+q%,m%жh[ Vm+@Jm%жh[ Vm+@JgJ{8_!0qK/Y2 $}qjqL I{{(˲EmvFݬ>UԹu DN r@"'9 DN r@"'9 DN r@"'9 DN r@zN =f|8N +`@V~XDi59 2r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 DN r@"'9 m'Z=u.^7۶l5O16U(rVMrN]!*T .FKU'b_rTqw=M N>\2-h9ԂpZ@DWg:;O_{JWs#+),n@t ]!\gBWt(!:BRoNϥ+N2p ]!ZNWҐ:FRh1$up5}+Dc!$:2RsDWX06Bb0 5WWQ2xteSC++=ct(91ҕS|NUYȀhZo8Fmo鐮 ]!گ Q] lzv,Ov,oqqչ\;mkz+k~??Ъ-_AӂhMoY#&mŌv}Vf\ovF[`F[hh%h(G[5CJ0b0tp` ZwkЉֺv dC+Dl Q}DWCW0c̀ [CWW F]!Z%NW^Zز<)B\ }+D)%ҕ1Z!6QWЛm~ PjF1ҕ5y> jlCWw8sWV^]!JMOAs >B4~]!ZvD=ҕܱ!%X G]!\BW #z9t%IΝ@璟*-pnE({jc^wzpC{gsvrDhzP6[گ^;ٞߜ|5yEY^vFW{]zvBM˜Р*c:q3ܫaBps,M7w7 t)_~ m>o?GlFi oϙ됄*Aä\bimtU= e?rۏ~l/8d:-𡃷Wi|=Y_緶ƟXP7O'WCwu ksYuu<ߧwf?$OumG[-%y+W%]JT&%( 㔲-DŊ>c*rvZ2_\[Km\8gtAsH¢4ܜUsNY=?EڜKtI !}(UvI#ιTEX& feu"i~wi%tGU¨a^smz=°KLOmOsf6,cay-/@)-uSPcNp}0?g@$,NJ6w!ո@}41^N͝trb|S{0E[_Vǟp/Nq\SηZ{Wκ 訷ttۂ '{ݶkz?hM/AֺCYn`p5.G~'T?q~\U\O1㻗!]Ӭ{@g}oޙ)J5JFyPoD-Ș OK3(fy[^ 3[P6AWJMXRAiNP"µ.)aTAPe^*a*]%8Y$PX~vUHRs5B'gWjm+tӧН.SnbuzR}7JݾrtU-`0fV[c"2g0gr 13&@d;乮E^xx(u . %j Z!t2X(#38_OlY o 6;{)ڌ3ye,FvY#|6y.ϛ*f6hcd0+Gt<:,f%M.){w]mÓR"bjH: e)DS >]u9# IV}Ĭj r.bd愳5Eɡm^1qvױQ)r0Q;] #PN&i #,qp,|ʢUqZTp6ؼMUkG=NB|Hmtg&h:@ӁRB;35%k&&%U1IR4ѧ{>gi& [ BAm2QXPR8'+sR^>1fU M25W,>=c^+ ƢVXA|Y"IN/Z\tʞ!z?{2PypP+KV4^1C礨d| R8CAk$w ̳RLEƦ"쌕LQU,3勷ф"mCˠcfiam3+d W,>0)bR3&ɂYgpgm\OafmG\Gýyŝ7E0Ze94t, jCd’5r]M[]Kk~OzND'9g1HSu͐4e-5Q*Q2cDRuVgR+-''Q 'ݾ˨C!IpjSr"hexRUQq4:&G'ı(gF6OƲ/Ӱ9;pκxͭX ˫IYW~+3LsZS(;@ז!%QJR)X]/:v=/zї묋<5ty [UlbqFR\Tr1|F[Tc+ƚ4З謇<5:Laqzk 8~ 8iOi֧l􀗤Bןl$b-6\[ugC:ZRjpU9*AHKQ:@}YD,GE`/EFn8(bsyYk%YYLj fYŒJGeSZnܟ޵ #a`?.rH;n1EjIʲ6~!E=F iTU.A2N2:ʜ$‚ NqADm JUD-KJhbDx3&@C2t39/S̓ EE\)AE" }{m"iShhH&mTha9 -tBkʘ'CXHN v!@G ƨMCu%weC&xS8l<d-9o$f(N3cs;eSgv ZϗF KndsA[Mh%2UF(Bk$$A <-A@w% s(1HHBePSݠB:ZաlKp*{ӶIOޮΒˆ2W+ ]~a87BGQOB1'q!׾ąZO'1Qi=;=Ϙ;m}6klV7:I36F7zڊhG[RlL`Skm} ܚaI}f\Kvi"mu%aҨC҅F7i:ZҸoHcj]g_$z~YWB|Ϟ0ap`i}p%RȥR5$i&b%2cd2IH) )"L\7J1&pC}L:;I5/0<&HH9MHX;9{T83KӼT6tWT#ķQIOv'/T}UĝxrlVRϟlk~dOuNL+Qܱg4ZNP!{%D'@W*TVՌٌC`TtTQ@HrV`n$pI5c5rY TӅ8TttQuᆢro낌{Jpw'..n}zWLf7bklWddP eV<he 4ɫ B-rJj(նjlb &CR)F 6mǜB1e[cLSոP+km?h>"" #/EVȪ)YB"S!2 `x4XWfm,HB&dYQ&KF$QzA:Y#dT' RV#g>lRa>ThZqF74 ixBeMh 'Bq˃"zH;Idi軺;XMd/\U"=0D2(M21^V*k¨F0.=@zqtuJj\r^"bב $/3夽L1${VG<0ŵuZ^<^<@P}(*Cq>< ۤ4'],eFOfG'rF=7D?Q#)"~6r6m1?W{iWGeQttkDj9=#~Xy''):8 (2j>z{d#`yy$DheU:%Yޘ R F L!39t&e)S |CZQݾC*4˺r >FB9AbBH^#K'`CB2.#8Okw'Ail֟},`¼ #K%|1-Ӆ: q:(Ҿ9ϦFBIP&qoJ& ]|d"JX>c-(KQꈠ1xd0L̵"dgh齳2 O ׭_8ItJnxswӶ XFPO)Xu72mfiGwS$0 B7blTYdҹWL?#7{5znw6K( s#hHq9 , \c >Ƹػ.J3\|Lơ' ~Z=F>?=jyrq+FeA8LҚ(^ F@LP.FeS/9$/pPxՠ)ko^͑yM>$u-9Lj2A[ʦ̐c)5[JqJI80]6pF)2zH̫C5r NIҔ&Nf~9hajRXGXC "Ee7 e_@d%TC=>h9-uQb&Yfb+El1{49_1GR2ޠ_IGH-\0M'NZUf6!Anėo@d9:EĨFXIZV!Njvu/`m>|y,V؎pgbcn1+:\O[IUG?v}nZ]y;i5S'^OhJH^;uĝMv(Q? ضULFE0\g'ױm?.}%LMb;7 Wp("6k]]q<?HݰS!s?ڶO\7|W\7'^oLG*`inI^t)?cFnOSՁO/揅^RBOgmg+7f31b]Ztwn5}b4ŏD|B^W}ݟ>i0k8^Z vU4Vcgfb.LBgenG>CQgun}v6]81ǻ5\_:mi <2I]}u霞 Ffz6'1vg,kb&nUX}Bk`{N]ʖM_4:i[yWqal 6' Qp%](P9 ^Œi,D*g܏P(dž6.zSO^5_~cT>mpʎCcwFEgsʽ6"ORiK}IA?m P@3gY..EL{laܦOUI$2Ge nq\dܒD$ AHF-ZevP'AKe"R짙iºWvV$GM2Y&\pN `l :;SR"f_NҴW#Z{bđH@X͕MA\2X-"A P YS y<+R^ds.qI+%x4(=sΑl@G4uBw>e>$fa~>ߍB?;.F?|\RKr4))y"M2]|?% *°"$Ly)"O2iDVTܤ)iﳠ}θ#c8eK&ήٹ5Rr`Har{X%֟˫_5iϵ*^?5yƹSonm~ₖ4!4wOjxqM߫#jmpK:b5햃Xm˶h0L$MS\ٻ6eU'I*&Ĺ`T5O[b gIi,`陝_O?bs{b|vOg!!bn7Ӌ(AQʨ J;y18j&z4ֿX48:p-^獕QhGRϏХSWd ={VMXXR!{{\rjCJ[p2AOdIEq |曯Ai%H_W7q[VTa`VjAu;sO$߽o޽>~yxt:>뗰~"T`. zm~[ V?]kk-u'|~uC7eRLug! tsns]N"U3IG@Q$"IZ#,!GhJ8 q^Inpcᒗuf 8uUcFh8pF" ^U$q;i\g|[[+XS֮uJV=xPyύA'jU4; cz0k/ew϶W9\!rAn)1-ᯠ_f=2Z(氥XPWNt(&T.4!ʋ '"팍]HZ;!i f9p"jn1ʴ ' I|g<"'_#8X[թUMjj mR2L`4bZJB)k, @Q Z3LG*Ez0k5f,`ZFLAPFSѲ9KNVPUAJ9οU}3LWw'z2j6%#U̫tf҄m]xͺ<)t "VUK&)Bw'>QX,ZT)pΫ[,o\Ņys;6J4-O=x= bA{ݻa~5_ r1 q;&U9TԨ9f3e)^G @#'CȭO)ֱP7frɴB0rw.?XIA$&2Is }nCi%z.0dME7(!ލVEB,@@ QDtx]*`%keb3ZEK=H(#BXYL^jʈhA #(H8 [Fﭑ$zZajaI?i*=%g:*جf~gk0y@s,#LzLCQ'mpq$k֭fPlozuqյV 03*>cd:V)t+'R1v.(rw2oV_~V(RO)prQT2(+qx˾? e-biNk(1pə2sޔ~$(h%$rݖDԛԸ ZyA+p!E*,ֈD"S`C\u+$tU"XnvUh[UVMW@e׻WL\ &0sBF\%r)qhURt +M`d{`"mWڇNJTNz +D0G[#JrնDPJՙڟҚKM` oZ 5`Ōn* շ#ؒPc,)߳8$Ň'H^v0V ڍ_tщWˋlX\TzSx9wxIyT鬦{pэPr\^O"rFfme&փR_W,anx44_eR^\qXt?ẃI&/(:N})0;(K՛g^d9|V{֜O-5ͫgy,![nu '1beP;/_,~JZUOdùZleNX[$`)ֈD"Z6^\%*2Ýzf"*)"qECor1 % o*)Ez!6+ W\E\%jtqz#"pZ|,cLWu GiY͒NyG7/Lzw*S.A;KMcwwU&f}χ?0mLm ed<\l* κw x)зxb룫(ǓPɝe/; $;*1Nb畹0py[ 9|iDzgşaLuۯ| Px4 Xb5G"m8l9J;9j(<].fg˳lvy6<].fg˳\7&g{`WFTꦶkÁ[^W֗qSH3}nDBY Zx>N}ZovÕ{PG7ѶP\X0VY‘#0Ff ''\F3j1@ӁVuR$\jq)0qZFe)A@Y@x8z̏ˡWX\],(ntsf<\k\¦3nrh< T uE#ЅҨh%a S}@iqV Ou0;7*5r1lA@4:&d<1âҌ;$8"` (O nqrAJ<dddKTI/1 HX~q' sP~~KQGs)m.euIޛ[*)M(b2@Q$_ %(9VQ0EwzS3Q8W^8y}W ,/:mqkno6D=&_q.VQTe:0 ワcʓ = Wٹ)I=K0ͧR^N_),+5, ROAMF{+xhO_UrkcΪǓy >fߪ;///>d̗ ?~NH` ^&A>x Ɠ{ 9=^guCwS~fz#0(J3f1ѤYbfz7VJ^GIsX>?2MW̒%xpVTgh"j?ܱ)CˁtquPLY2)| ]ҵ CRWCۙ&GLJ^G/ãcL?^;x/v՛K^"5Ooﺊhk*L׭emyN?I1ם)(eT Ldz`rc,H׵{$$쌣z(E$WFgRHPwι(\\|KWY5[Z@ۘUgsۻƶΆtr癡Vu;OåJUږ̗;jgۂ+OрmP9 \ҖWw!'3St%U:Zˑg푵WTA @(|+'C:w^*EkvF۹d 2qXl +ݥ?۱pk<|N#r?^<8Xh+{Ei6/H@WcCU deg-j!abF/^ ~*竤|dAfVX=z}3xT|l&l"#wS.oΗE2XOF.?EˆrQPEAz`cS^t(Pb4Qj̟eoX~%wr9VmafHnxv _~?fӘ͗~^-m}c SYH,ކ[lkU?Gɥ+X( grݗ9=.9șJ[+?.\əazszGwGvXY\ /׍{iDw*UN=$b awy  awh16@Ўe 8G|YfeU٪9 RpA)8uQog]֟7:jiy=#e;gcn1&azz8P1ls.}^8fU!eaGmN $'BЪ`PdR iTk"3+Hg?|t{suYt7䳽8:P\8>_ŧB 1 'EqQ9G[ h)c)ʖz޲7EJ9_P@z) faE5Y[AsˉsbT.\tկ&G11|r%M mlv%dG:`N8)v<}v1HeYPmJ Zw"`Ʀu9Lm4oZ Áo3fk-}qH</&T~@;@_}c}V`K4  9MK$ⳬbhwQcB1]KK49X)טRɘ`*$cفuucMsBv ֤'0rڢMDhkT5 4&FkbrѪ852Zn7L[!g}P'ʐP,YS<@)d ŶE-; u0#M2K1"ȱ*Lu@;[տGW9_|ܥrlvtL^ܠw@#%-/"ķy {%h=M돢+G}~zjeS"ֳcDκ೅r{S:(J[#9~ZݠV_m{?* lkK 1CNZ,"M.jl`c.w躤+*>ۂJ1DlT,tAu(!WeGw[ \`~vcIHmTaѨ8/MEKBu3KGc9-?tAkʔ-*U/EUMB,$?C2:F}@r="| cu'צ< Bۛl6~jGЂp3c`msē[~bnI!cm@'T r8drm5ÞBWE^}g[ZYbTT)BI1)8a*Qo zcqzqcT-Xt<碛ゞliл,)]s{g}=kwq[#pu &2InR&= Ag|?r 76,l7?v q\ӅZIC.dħcѡPI$g~UHOQ`dL5+\Qд&.LVPm̜391gEAP/x_I(Cak b:JT0$VcEkˤ3&B*}+os.N}6wmg0}Ngzkv̭V\оn*aFŘlXZ[ॐ D ɨbJu|v}R!cώߨXzzs@" ,N[5.ϧd;zdd>Jk}N`kR|}8nfbߙy(EUj ) h&GPs9..K.!!V3sV5j#G$,BS6"3`䵷XSz{n0V iƹ8“µD_2`yp10k~VV$n.gcD, [NMtM-m0Li`[âF@C#{HFٜ5mZ쁂IX sJ=v7q{|3'\v78k^;^{D;5ޔ,4^c 6G%HK rQzuGKbRj]l kM5m0R6 ĎTR].&~ -}zI9Ƴ(~UG\_oGc[xYu@< :f>v':)I8x;_ [/)jZLɔJ_}K֎=X ɥw-A7q%TUgae\%E//~qg M(!HɄ꒳  F>0jBڅؒwgS`K;?4uVؖy^H=m3e[nI^Q,RH(YZSe?>z}35uH{DYk 8@D7D ~ZȵΎiH{ j5qny;|ǧ2W2Y9zfSZ0Xn_};Q.L=u7clySEН,$oLh|QUڊ5iQo@50uTx#U&<v/Qǔb<2@xomT&9(ȹ)zfO5ɾ{z9#y# 㬱d8d \.J7j_]. QBK4N<bd&GfTXO:ka^{X=b-Bٻ6%U#}0nؼl`B!dxV|!)rhȶ,qzUuu3f4j|s\1Tk%i|>~iK1OJy6d 6d)=4} l & {y \>xٲi˹ep7~45+y#~kѼ's$iJhc<96+۱%d.ʆ"հ--vԓc:KB)HyKet1h?$A,)Q6b,#a0ಆH&& d0bVG~0; lz)SؽghZ[FGw @<&ɢ'c̏'z`m3&a "GLEnlaf$BK!'V`ç RFjRGZZ%"P17B;r&2-g@V)ٳǞ SH4p`NJ7bi1g;J1Cklqi=& H; WS\4'fVMɶSWPb}5ۛ)yN 7g>oM>sF(z0F#A"p0U)TEMzl>:c꼷L`{"ukEjOa8`rQ]J?F?Ϛ@4%EDֆ[R%Roz3yQe2N.6ڼrh_*zFUv[U'҉'? WpYRu$/^~JQ37-&hMF W/X2d۶񜫊o򟫊5xѥW浾3t%ý}p-yr|Y*5f$2ןMR0_9)z]-`᫺s\CSz֯~MmZZωnAV0{ƛMpaL+5D WLc3~8} oGdx_[a9OJ|&>qۦkf$TZ7*3_߹v(6lszZ/ѸOUT'C4 PUjg^1oUcc&2Qe@Ejjk]ʒNhÈiv5T#Wq~PwRQ7Xae 8!;mTcogʀO-ChiuƖVlcr=1;tB) x;M\+Db.:"ҥH<D-p^H;(lGM> Xƒ6vM!$5\e |(ҏIݥ8`oWSiQOrӇq ^rÜ9Q [cId+LwW v]+=k>qRYJi# 1͵Ibi[0Xn$g 6p m7E7JA- "A@ J))$mH9!C)tL_|7 إ4ܣ7鷏h[fC`{;2GaEC: u$# cf .kyp /_wgeg?'ߙOCу {L aǔ#}6?ѱl)Q_ DJVșPg*B~`J <{Cli7S|X^Վ l !jǡy.ʏ_6zcV7POT4+l<6+if) 2([XN~8/ YqS ҈JW\NNE\%jzM><'Vjh &uD9hkcc3aL7Hv3nղmS.u0B̀`fܥHG%j):v,Qf;QWhɈ+ 1%%j_yTrމ/P\QrsJ+v2*OF\ǶD咯F'q('$ZɈD.=e0Q{cURN\}#M?INK*ldƓ,aAYWrzqCv4Hk:@^\3I}9# **:|$&t?p)2Nj& $6=s˗Y“oԤ孄+MIH|h0ڊH2AIRPGY %(^n/uKRw{3o/=Pj тOd%w-.7 aT|%V-a}(hPBG%ODaO(`_-|.v'mZO'z1G ^U褜xtxJ1Pй1Sr鬡;f#}S}mNr: ժqiՅX:zGZת=BmU!UX-zg՘IDk1hFs+%e5rZ)"TI)+tVF&S1O-%VS5NxF0]29k)(vM*iVh%e`Ingd *mEڜciOgp TSlB&'y'(\څw+tEhٯ{$\eհw)eIPL@IU|P`NS5~-@uC;enʹD?&%^Kh$㚌^MEATÑlBY-D=OP<%޽ Wt/\yΦ+No m)CcLٝgAsT,׃~45Y:f|'|6ZО2!!-*ܐET$ҩm%t eK8s= r-iezD]>2 9#ĝveuFt^n1v)GQ&c\}Tp9ꚗJ_pM}77yț {VoHWtvK.LfWu=4`ۄ^}#C:Xt/t05hv[$|4 i/`n$oãWIAA!y0ERksJ/lqKc4bD.(@ QDtx]*`%ke"qQVmqK=H(#BXYL^jʈhA #(H8 [ŵFvٷ5.( ƻ.`Sе1wNq^}"a(#Kb&:iC( Xn53`fҥk%ug1^π.-Bz)i[ḱ[A\5/Sm 9S:K$Հڜz!=*(?n;GHmiqmM5X0vS,##Ga M%Ձ J4rr{̨bq ZŤHXG8LI8FҎY | \K,hx8zTH2.λXDM"\)0^ @6jc"Z(pN8}yp}74{\Rppfg/M Dl6HFd lb$beݥB&#?_]&$:{lj6"HAG2(C(bWjp`$$]aX22`)"#t~#s3w>|w:+`P&vE!&h{r8]UE5|t|i%ɐpL!HQ}]VWHU꣜²R\cҡй۪4Ykc`EJUҧtVO50S|]?g711qwcorl^풁$7on`>0xjhIavl[z infOٻ8ndW~ٗw '{ dq"Ҍ2#Yq?Ş.H %~=Vw>~U Mg=FpUlzqctsz)q6{J݌.~ZHl޷{9]0~I_lpݏJqc=?HV&?4{Hu!|}%To8͓ʄl~龖V?).[uIy]nz׫~qC[2߯kVmTqQ!CFer[29.sM|w7?s;I4blZS-X$dӀ (` 䂌1ecv8!9֕^ h;<ۋ )%cL(ghקA=I:mutrzZ\7?lN\5;{D+ JYY >d5Mpkh(;N m½R^^{0wlCM;:Cbc>EC7]De(BxOH6c|6w-.\zg+9t_|A OpÊkȒв  VgrY}CÓ$,6Q{t$\Mڒ=q˝x`p{WK,vYaw;GoN=z,0So+S[ՃoX4b :+}eR-V>m'TQhh:-FY!S4rNNN<*%Ht{C>Eu <}HYeI8ƽv9U4ׇ)ʊR-.-/ !)5G3P%rւE)TA,|@ƈP>Y4Bh!g&@\<,!TLpU%R(e.1,ފRv6ٍr2e@J3+xQHPMYI ~`Bh1.Ɯu/>ҿ}_ޱ ygW=bV 28/$:fi){vy@g:kߞ9eR:f1Cء3j 7PJz$EqR֢ "%0 s" JoRȦ))cb[+$ƨ$ f #F?% !;Yd3,Jϔ""`cRL;XH_Hm]z nef/=oKˋ샗gG,6b|qdGĎ!vʏ -W>kqƘda E3Z%)(H!IwЄ՚"Nk5^Mu%(q#wQAϭs9D6Wv^sHY<9{GӾm`]3&c۫_NOgg-Η*u跧>!~U 4qͼ+wk ?wbm&E¥.,ԍKHR~+([:AEtZkL 1fĒ2flk'[OVe,VuIҕ~?TGt̔ʇNIYFJu?A`ݰ2v|b,6n4%B}VXӳd!ڈFUeĬ|.*g)`nJk "eQl-PC(FbJLJ$2@* N#o&[%% ۚ=!ԍZ4vw7G^2OH1bzכṢpDvxt4vo{ۀ~D :iLLERS(D3~EM@|f6-JPIFL ͙<* R%Ix4եY6|@dRL L8QqVed+qn'8dzÃ4 晗EaBK AT Wrx]VVrJ ¤B"[RQS)WK3V aR2E:2VV>?:FGՔʰmaG>9TVB..A6t@.jSӻT(mc,|7FR&θHk HAWd1ew^K21n&b萘]Ft|Ao><~ʫ.e%)(R`DH",c(Q+6FՠE@dAN6nUwQ^ rF, YgV5Y,&Pb1X}<$ /u D94ճ(X!y0`Λ*( _7#\JX0 ypItc=k&׫?f_&*xɺ֋9H~Y7}qmf\}®t/~E/%_G_N ?>#_f, Q~K ~]OJ ͒Q&HQ׳~$^ZF*|@e #RQ]K/lvfZ>c-;/[;祣b _bW3?=Xqsr:.2MtMjѱ _{uILӃZ G^Y7g ۋ|"L|.%o[>EgVm8W(YPÔczLҧ(ȱs_Df(;r'β9EMY9e, ֲ:g#ZU{xOk_[3EAFj)"+"%X0EW>Ĺc"ƍ=ݧbĺO >[:F5I{K?@"hD(&GJBT;!,Pm?jJұ)STI&2J╶H63Hۏ^$[4^gI ǗlKJZv6=ہmi1^y}][UyNV\njq}]sa  "Qb'%Kv8`X҆k9GS3h %LlPHJ1Sv(5*FfܡfH3 ͌`!XxXie['s,>.ҪyAgg/`ȌPd^'`P/UĺYl)FX#C5bVB#**lj3L*d_ض2Lp%QC~Ĺq6L̡vٱ+jƨF3^%mbYZ,AAU|Jga0 KNx4:\Mv1b U-dfҰ4yƚL"9EVC"bHlT#ds;F}{2?Tq0j~섈-q{#"-Xi'r˄&^!fctHX[D fґe{1hQ fLlP:rTj]&)Lk6vD<{GqqѭkHfdW\qQ8⍑>@ ʃ#Q&>L|b2m0zwQjj7<DmI8~Srǐ>b 1XYY 13@D:֔mmQ`xa[0gj{ȑ_iavb 0| IƺȒGd߯Ւ/ղ5N[VŇU"qʖiƂ-BJH[>6t TKo$Ș =ֻs/jwkQ}}:lr]TBKkbef6DpFm!˜NJP%F]䗾2\4^iX\`j"Edz8)P'ɏ0'XO5ݖ~7L>L7Ԉ]릿>o8 ߯mn< ZJxק 7;E^VU@LjR?k#5=ht֑N&hZ:SU" 7fU=oo5~seά $,83>Q|żODfĈC\+)s8J&>x6?tדRu/j|G~C0ZS`xX{X0zϫ~ӗEq槏 FųH kU1kexň!')*H wGJ%۩F#,/@ C Z's bh\&*Ҹgp)1tkY P-J TUBD `ff Zs&aprc0<՞JCYr@S<gL| [*\AζP1}B?GfDpl^@&W3*W_9Pc Isux~x" j=h(D@ilG@ " Oih37jP֙MqFdz1 zʾg_N}/^6fz,|wXGW3!\o6Eon\v9y/y,/W{v}" LP57V:-~\~UV4O+?K:juX=d$SxAKO^+^ef]{^ zs6^:VhIFtu>tp ˅}+@@W_]-^΍osgToVG#em}Lޟ߶Z\uSO /◟}Y̚"@k%ens*8#w}Vk[q`WuXt7\UVM8A_+iѸeRw{1wয়΍Md4hv$H+NŸN;PaT?O*vPe10Vm- JhG£~o:ww梒1l 95̶e3"Zn>"Jg[&hFtU2Z8FЁ]qi"VBWVXw 1ҕP]`Ѯ.g+D@WGHWRqW ,싮.1(:JRJi.3+Y6tp ͅmA@i(J+K)ˈ0gBW,Ii Q9ҕTVΈfD\ ^x:]j@WCWV rrFʳ+˲1-}+D9]}Etz}Bhf`&O);[c2Tl܌ҡZQ{x#`A &Nh%;ϮJEzU4~JS6{?jj,VJx{OS#԰ePm"%ܗYUZR'&@L4jj%sh*R*(qzj +m-6hI+"+'Z29-N>Exu{oaUUN@=79׿?ILu tZ IhJ+q4NRs] l$] le66 µ4 n JNmfN)=t\A7B]!]q#IB&0µx-'+DI@WGHWh*mFt%0T'B>ϥ瞼NWR 1ҕ4h]`Kt6tp̅-=]!J:BR0Үɇ͆!+DIJ[!ψ$4BBWV1ҕÄɆ.=PA:F'Ԍ K³+˲YD@te:Tvն]oN>Y= z…Z26nPo$h1[qSH!;-ۋ/%kup-@tCU%2L- ZƮ]fDWXl J ]!ZNWm+NW2#Bl Z ]ZI{]!J.:BXv%4R\ j"NWȁ$ [1p΅$+Ewˆh5 Q42͆.e]J S,+W;BfCWWe2h ;]!J:Bp1•2B]oO*jvl&O섖*[g7gQl]R{= SFN'Q:[1ܛj mS-ԕ#-4}̄M6[=+IL4LHzVq6l%An-6\l3Dmm3D)`mƸ<'W&\ rwBbc+.$"#BBgCWW\ wVdDWr ]!\1hM]!]Vm融VDCWf]ZH Q*1ҕ sJ+{3yt(v@W_\l l+D+z]!JI:B/w_jO9٬ —~eQA:FЬe4 wBjXjJ^•n-.A xcOISB)6mRmũ8R9pp|s7%ٞeީҭs zNꡊS1+tR9 TIsm0;YG2"Z>"ʞy̶T EYNt&l *{=;ig[:45=~4rJ_^Reex5&ZjAj-`ۋ?~F(i̽sΠj,k] %,yoE(͟/jΦ20+Rwt}KiΏu% BS]>^Afn\˺V*%$_>.™ BJ^.w󆷳-l8|/6;Czuwv5n] F|MŲ6Q\T}r4ooYu}GK^GMeۉe("Fpܜwh 㰟;s:gG:stN=CΡm?tq)3n^MJ`=?sLSoc$S9Ι4)RexpV(N|+Z)qTUq%ϋk|Fn9L'/?5JttV&!~ZNhEI+I^n09u W+/# *-#iT];- j"'zM 8'.K];gC@E$(WS;j656xػ(*c"Xsf Zhc.ORPAa k(aF}1AB!jc K܇ :R!"eP`u<w;+Ҟhid. r5=?JIR q$-K2d*MJF-THH$qw!n!KAJ`nL) :f6W,ATT"TyA+ hL%`[}Æ*ӠAPw!9"\0HaQt0 M10$w!%Sev`Ct?#&&c*B b:X#iه*ެ21ڸJgZ{8_!M[~0$0HeAPO[Ytb }mJ mXፁآ橪[U -IQ1'B&4AO/Ƨ\x AҬ*ҋqZJhH%%Ɋ] s eNJ}:U2)Di;z7Eע$S&mI!a##+$V 3_' H/"BU>) X )dgVmJ9b2m6e!j$ԓŢKS}h5ST:8S׾"fnJ}Q$\; Ht=+ )D!; D{j%x!,;v4#H2P4@ZE l WSN " h[UAEE'{'0O4v/uh;n2Qh?i(QLnB!VˮRWHme±&6V:w. 57)B̆f;uF5`keB`T^# L%ם`>%XlXkY"0PVPB ׎]SPP&V*¤T) `; n!\"U0VjVa   ʄWws42N d> A2&d]g(M DBΦNRȠU %;.j!ՠPwVrEdܠQ ; @rJBB("2m ؍ttg-JC(]eԭ9+ƒ s<&`&b1S|bDJpPgSO nrPf̤cB8*VZ hM e*3[!(@Hq;A(x3iAQ 2#}@PSQzX]T;jF'U]%їT=#i̼Bիf !ѿ&伖bZvJDA Ք`d\D d`niXEjE},B\FФBufpA#k^̈KQE;fD$' GcE( /}0bm;/ jݮ9-?Ԃ[Uе3[Xw ڦL[oT^ҧMJw%SdIW=$ B+X(ITZl?NڳF& h%A J"-Ei*5.--U{9 i ߤn$BG 4/^m:hIYT0dkR@n㠝;[t eu>`t1˴er$Y:ˢj0fU:9 6IҢGJ${=Ilf1hmT%8O)'ޮ Ę@9HѰ-'' t&% xKd]ж+9]PnDVܢzI,WR SAA ,3 R4Č,mAz |"2PAzzT [bcW,XB|D4cҠNn )қDyŰ aL+JQYTPcD&7cQGQ1abY;p`\t%i#*`hAgМۦwZ,nF+5֤Y 6)jP%Hhͤyfj)څ'е=DT赫>@Mf ScUwUA`('` aвMf@ 2)Ziq#f0YJFD9ikJ’#iVAo:^#@.* w6&TS.\oH1ˡ옅jI I贛Pb驓:HҚlC9oU hD;S`]pBL0H5寷zFo.v/n^.כwIte0 &٦aXJXCqٳ;ﴂUz0H\R @-v\믿>?io>.r#)6R\A-7itqA;irbӤ'7@a8r}~d;Fo5.Wk\zzuvz*"GESi[+bur?Sogw_nO жk绶^.ƱŸmϤ<7@A?S~"jܜ@f|@ϞmB- ; +v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b';e@Tә@aL8|0#:PZ>0(@JF! N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@:^'"|"> QHY#; D L v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; $Ȉ(l@@;vIFv@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; tk=jnɳKjJ׷7]Z;EY\?|cq KlK@"qKz4s:q׹Ptute#+Άs+BG?1tE(}d:B A1'uE|]\7{W6+B-UQWl=';NW#^>w;wxmH0G!Cy (blmq㖸M WB3 Cųbq\oZћA3z!>&mA~bl}J5/ߎs(^b7 dGpEX^bOOnlar>/ӛV~VKfj՛a}<]ڋGo.FUsgi̫7-Wm} 5!^s-~ 1|-*']"ׯ˛ElYDWzWOObVuޚmA Jَ,Y|bZ膘jעY &$;@jwe?Gطy~h#-خMGm&6ԡQ:gDW|*=":]JchǶ3+l1"Q:]c`^8\ u=bP:tutebN̈8X9"~6FwtQj#+:L Z?"^:]tutΈX7"v6 PtutD~N0"j6tEhPZ^iej'6̅j+J}Ы'Bj{ldmCW{<0u+tCOõ2Oveۣޙn]7/ Xbl^ԟ܋=lcD0n[3СC8d-bT-<Ɂzg/Gq|3ɫ˛:7B&u5/\9QK6(VUzTEjG]㏐ٚ1lI=us~ 2;D_׋|b{z^z֠T^I&"'_a0\Tru3O~dR݌4GigstָCt:tG锤fDW섞 ]\iBWV| J(| BWZ {5"mcYBypBWF)8Oh<ͩLWΉpO1h<"VTHy|qyb|t/G'xe/Zח?_m?k_~IԬ|/v0!RM7U zlḫT== ΍ѳYzȿׯOg3zU޺ozz7' i6ڮ'}f櫴NM4sO%ۣUzչZZJ>G'cZ+ !^ŋu@~Ӆt~J䆩rr}&ӓOOnVwN(!]s~ ׇ_+t9.bIU^|<̵Ng+oc<}~A\JB tr%͛>|.Hy/ǘFj-?%..]o/뗷޷wT{~[Ot?l/(>ߘͶܻs3ĴZr~xs9&g{8W;\Fzocp&O05d8l%wyPCCRRl0EMtWtUYޅu"ޤ]Sm'wsfgۚZąBwٌcminWέ 4e}>Y~PO?][yi#\(wp)BSϘ8P=x:FY^1p:@j<1"gU]">ܲ9@qq ׵0սYTd|:)n =!#e(K<Kew$WqY>o~(?^a:wu{Ĩ~͵cUvQ\U}oXِ~=,JJ[5RJaICbR~1_wKnq}ɚ`BZh;k@}PU7g+@oCjmb hrsLx% 4r25֮|o#GzԽ][W~pq\ 6|a5R@ϕMiKS{"3#G PO z"C4ۂ$%cUV~/,z zzOI5Ezo8f|OasqOca {$ie %+ d0bVG:} &R/ _xd!RдԔ1тFQ4`pHOLl7po`ڀ]@[^_X޻D׽qmݎ %٢ޏ񜷇W֪)>/EyRE`k\Z0q@ P Q"`f|R85 YXQ(0,1G I"' {η)䩸.ɻދBs=$;fSyA=)0}(?sF(zhYEQoa:}QWlt>\gwݮlr5Y_?IqyT{]6[rjb~S™|:6`獣y(gm3,#YYE? J[:~^k$n{6u|skeDU\Bf<,dn:w|~F?HxDWqߚCʣ|r|4[ߍ*Vi0e7Vrli}iָ?,ːWfPn&6t'y=TK^](t.sO9T?TKѭZ} g:Aj{}C*kK8cmjl~8jo527Y@SEXl Ewvj ۟_yT zw ?G;u{ylCX:7!C7k&gX3LƫZk4m'ȫc =ȵta΃,pMphpR*,%0B/@eX,g~A6,@zН0Y*}qb޽E˪٥);aq;Y'rp9V=ƒz#7|7pDJN}㏃Jr͙a.ReQnjΥyGEߣFRqE%{-CQE[S=GTC'NJ i厲\@(bϑ՞R!F4p$CppG K25(n_ݎR^ߴFNg(훺p RLK"hx _D2.+!etC€JAlM R6شwL^33ֲ WAܕՇOJ)]ZRL-JɒPf9r]Ky5(§j>zzHΎJvZxLGҴ<,UŨ6g&!ak&ˍLiP3(,3Gm`?oÖ4WpEӲx%EưLe9?/ݠR2L96N OpQեٻ-ÁPMqb EӞQpc^ *g2\dʹxEv$As>aA[ZvK.(]tf]?KL R 9Uduḥ"ot$%A(yb[t?LWWiN M8 r>@ރ侩UjKďĎkt6M}*\ak7y/jW=Qk4'~Ij=mn[ۛY6a)ANT|fZ?2v+ų{69Ƶj;w"m]{Cv8s=̧uѯg덮-:lb{f;u*=$xdrqItHdi{"KIyg7~&m's$"{cpӹOO>сMpË::p[S0c=ȴ̓?:?L߿xZ.a *^4' 6gZ5^)OA.[>k.:IЖ&dڧhGr7 ѫ!EO<\$&>86^byzuCQ3H.(@ QDtxPDI21VX|FI0>>~ oVsӥ;YFXj<$6~`QAan53`9Q&p>f2ʺ_!mI[o9ڞƨX][Ls,?)H3}ny:Bj@mNSB9HGTR,tأtI-!'nj!iPLu+{C Oc*5(K)µ ςȳ1`XgM.vUNEns<6Uv~,hzӎV%4*ZI()h'b$DZ9؊Z +egcncioG)T$N%3,*͸Jb,#FԤ^2_+|ݗ7B7Ƨgփ*p%IhAXJQG5Pd4( Q`Ii?/o wZjk:HkX{"j*hcJBR+'Xb : mH2QgDj՘I$+KHSm RXDY c M Z ARLW3O|mb \t,|C_b$c{*%f(g~slꤻD?q\Kty>d); dA#L9wRP;3e\: 7eo%]j:%:Y3MwVU0Vjҋo ]YoG+ ˔wF0^ٻ2v0F` dddJQCi}#!6I7dQ,ZU]QߗǏ;{|^U%S_H.mq8W_Y:]D9' Qh8=;߉~槗?7o7W?\'enxkz+_?_^qkMn/&Şq.1̆HaNTf}B'+m+ ^ٮ t^ᾃ/)lD'/B^lV9M5Z-Kpd(拭lȅXU&bK-Q~|Ԇ쉳_ēvw8bv7ۉ< ,MjdxF ZPa^?MAry+gZ hPyQ>gc^ `* ӕh"ȃ<2iU- Z .6BuH腩hʻq^e!ESJVttOB|ʄA[. g@D@AM7qnгbSoc.sjpunK,*)וz_;9Nz͝#Ioʯ"h}p9ptqHB^( 38x |B㒟@m@ݼy$58箥:Bj?6sj2 ǃ|(|Fcp6^ T8-.JsNe>Nޕy6=2_+n6;mys?'߽ic{'{_엥v^yt(7_ZvO<7!DNoTa"6ao1qH 6֊߰ [Xʚ4@(&lFM#1C셩}^X^ĈZ >Ne]Ә 9 ?ޞly 9jR1z0hHt*|3NsC+RňI&GQfU"ȹ\rAZSkogk`R{ڮV.{CCw35n?'{W(QUF&c >@T j2AYb5!KgԾׂZĤSo aB6PK1sXKlĹނ/Y߀Mis־6{W b(| f Jyfu>pS|L:gqhJE`7łUCNcA]=S+(B)וI%b1АJ>e 2bddT9/puLCrkkMncy@.*3 'P]5-F.L66"J~ӥ{5KֺAr-"/q5j δi(CMꐳ:lbSCKBߴзE7-q¾rBbrs [2smV%$Z6ɔ}wUlwO.*(J1;=D{Ǹ8CDŽ4jeAgP-݈0߻;4*7j"zY>V-NX5>%`0oRrY,YdgUIaP-<8,)ꨭmnimT],)1h ܬhpm-*1{Jl!gS=8JvMܮv߯PT9=. DM ޱm@f8DQt^,} f6q 4Oߊ/ CEdl&ַ=]vDVc.+mLē&HqXH69z1c^\om`ijxtnmn[((2nr]a9:8Atbl-nG7}J}ӲU r圢 w ^َк@ȶ;j & ^[ꙃ+]B  KMP̩q8 iyɥiyؐ?~utӐm#tAĚsٻa~ɰ I HM&ؤ &w_CEp/lC$J-\P.e6R1s.b&OI|FT'l)3%{΁ )W,P z/NqԭrQEyKxKit^R$Ԃ"^Uu&fBgMF6_G6X7xfԙB"# P)LS5TrB֋0ZJ널(jNE4  P1_g'.d#1hN,FA~87o]'F <`<䴛lINϢtbJ{G6Dqo3Wl.:y4.O?\?v:2|?Pd@X 8ȩȜ\Wbh]=cX+ j#G$B"NVb᠃MZ{'kroM$w*|a7/ީ/\KԞ! 닓RN/5~>ˊ}C?9<=NHf2*f(Dtd]&Q8{>PA$^b,MbhcnqXKU9BĹccݴc[6N^{BWj,%֬}ȾFKR"kcY9ʗ ΡECL@LJ CEg[.*SL+CF|HճBKLs?VH`Ksc+_(#oGlu&.9j vq:jl2 gЊcSmc}="VoPE-s3bB(C R \[p)j ºsGBcK/ޅ_M`b/l۩n=|xT>e~1Q,*_7E?>PZ~\ͲĪ?>;=&UI+f48H~ 5o5"1zcܼ?=^!=hjt\ѻ,BslWvQ /Ń\p60$.4/ʭi.H:.kh3_ QQ_CȺt%e rECgmZElTXeLp=ܓ5apc> ~lQ ]_ȫtʫؑC1?|;e_-PJ\=_vuҮv#dAYeF%#|9'Κ}s0?ʁ9dm{U$ʹ,tyyܾń%kׯ+/x&7,s84s3׊V7X l$9:>;\8r_~-<D*no&^i^,>n&8|U'oٻ6r$W|;MH`d.9,63X`g`YHr&[ے-[m[A,UŧŪ{OUJxF\>uU򹨫B-ྫBԝfl9p(++pK J]-Vjcu 䬇 &{LQ nO_~G#&|GU\T#RcJ.}~+$jn[jNjjn 4jnT^iNM{Q,&ˬ }Y7#Яx@5tGo4Lv~wɴ[|8^thExKg,'L>R lЯwW:-E8u]9* gMNgg[e:p7*;`?+Zͬ_Igue]̕`ֺwY±>n$ۈ,$XYqU0F8NKjiȉ]Tt6*eΐDqe:[ q汚 ᕍ*J%Z rQ J*MdWJ5j^̧.w/Z^D)=Rp#zUMHeNQZV A'x򭆰?XsZ~sE3ua,4.*zZyh3RAȆqddLKr,/.Cn9O3X/h 0gj̋_[r0S41Vi-y`~f| |G܋`.E @ &L1.z1r sL8R:LL]!X4p5qƗD:&,#Ze)yM9>ZBs$~]Stoxqѯ;oΧͅ:1B*\All.vAB6|Zx{솋 -5ֶ$N75#67#fV bLQh(|4\zlڟ99nnuMn+6;AS:)4Ұ"q4'rxf.u/lj Sc/o$`\\K]BB,E_Ƨ3yu: ]ĩκnB옄Oѻ?W?ͻG\7e-WI;j5_o޾a6M4jߦ]f7v%e4*UKhTPF2$-n'xEY`d$Y˔Sٓ>OPܹ *\g'=a\ZzA7´qɸ`ldo8e4aत*I`djJ>0"Xξ^-?n޻;>J˗aU4;#E-_õW㯿6Ίf%!q_M% T4YHU_%~Y>V|ehˊ[mdK=b4Sٔ Zi+:a{f%!ĔYdFXn#"^ hqve9zd*=IB˭s;L"8W s59.:xyX <;y.صO;y"NǫXsTR'Hf4,ZN^̑]&$YVAo߂Naӛ#:eX.KYξy"uVlmLMWŴqj)J!F)r 6wÜPf](o}5bt^oeX<tY)e[CxE*ْ6PAJ%Nȍ F9y*M"ڸ>QfN[%Xrۅ`"g=3O@[pVsr8 'e]([jϬ>RkSChꛝ >^%46wt0mD4F*VlL^Et0TܷR{>LF I~C"3KÓHM@< S1%VYf)ǯ'ucp[.e"-fų-iFFA$kJZ7oGA*7ظ.}އeZGg}ZtJ&I|2>|*i`eqS N4٧H$c*cY"K! xyBOG/ѹ~#0G KSxiŢ(&s` __POv}$9Dس N8vYf옲#z+Ik/D_@YEMFNEM4Q,psV><=M_ %.8R㴬\n%L zÓ?hL]$0;Wn\.nσT\oMt.fm|1ӢWnȒ>/T֖^CۚIk)\1kn~uy!qo+箢+zy=}i8[^x7fy~І[M{ˇw[|ͫ9Mohtwo - ϓB.~g_Mф+Jo8`yn{|gvw.6\nuU?Li{"O uOe̺hWg7LJKepנT*$Wbe ~"r|qל/c{9k d^Ny9I1ag2SVG,ZٚV}q]&l/^1g'5p^lH6b;&I<^K!1^&i S`Ĕ.dVN)n][+G,V9vܠo@WrgKqTt_e-6Kmngtٮ#[v|84}s:/v# HOi7rc~a7j}XjwwRD鴩/ԕcDr@0%r&[c^LW5v]. #O}rv{$yٶ! /8F So=L}e.9;p.+i,mh nSU{^1*Fl"#4بK4VvA2' uJ`,)e1+Ŷ]:l#,0B4.uDZ@!rcJ0 /:Fo٨nۍ1}6Ş棗/< e=1@4&}e,:핱D%2@1 мPa](LF] cdldlfŻJDO k/:n k"zgP{D$1Ebk>˜JZj}Bkܕ]KvwwGl Iַd /tc^ Ȝu0QLd֊ u^`i JS)-u$d}22f@%0$5'˃ky9C ٵSM֗~n;!De;gSjrCPYdR*6sV23 L;cs)Td!tۚ5= %gt^ZdzYAHƳ"Adº`xT  ccy#FB[Ʃč,;oH&aٌ!:lRBO5rcvNqjbrᴒJq -FҭHep"9zDC zWEXecЎjva^$# 7'h&Md=mD?Y$&Ol0{ϦBQؠ A5*PFsDcCQ8G{U|J2}V%9232EGAt& ԘQGSoMp w,/O mhYyt6x9iA₈djA6ӌ\UAttpI)MY z^:ƌ d.,5-j֝*ȱy}H)sHd;rM$4%)q<0B+!7 9iAyɍUL1݄ Fizd%}5{Z\Q޹G?Y #β\,Fl=<@PV4ZLyJZ!IԟG:O)@?i$Sم9U$^nR@p2֝{HTZְ9TYԺyã-iIiysx#-#vm]cώ ):;s8/ R[|Ĝ(jfa jBdD$Æ&D'/$=$m^kϠn|]|v6sW''e)9m"\{<_~C ,P,:rQǼk|!ҠэWgiօ-?Wn0m軷^h|5|)}_W4׮;PՕ]G:hW8u]4x5[ ډER+ :u*)Q*)O@H/OӬ^(3kQMsS@(+Vzs,gY;YvI-'όtN=z":1BJʐ_~RRlFiDoG0;;]2KwD,QOZgERPF8RV;KZbNK`.`եM5/&D!!Z&\?`]ٲ(BOXc̵VMBB#ң-˚ӦX!|>j2mGXϙ]S'Z 񌦘ju٭xԜ^O,i2&p|< $szJ.8f,0o 0jS%G,Ȓut.fYJW~ %(YXJYrb D&s%8CV d+jݹI\L_I3w}ݑk u#t3.byn}$ݽC/@ML9cbjkLZBr Iɹ6LdL2c"cn>^x$6y!].' BHR;Q<d% !Oem2#L)R*[պsOV>SzG&ܿ>c ?&)RfrYA CP,C`;J&RF8*y! >&ϚO@D $VH-S7[kuiO'_Ӥ=6 $߮&'{Ye%5&-YZz}xEI&gq16\~bq ^5ώ#C{T0u,d`VƁS)D2CJrJ}SmA])U1\)fA0&4H\+Xk#cFz\VcBcAp٣lŒ7l.NMcȻ~ W#p+22A2+d 2UrF\9\e%5Ruc~{6Ѩ`ZfY(79̢uc~kug7bWLcAjұ/jʨ{V1Vy(28FVL" dUFã)Ց9ĺx#" !WdEG,a &HECđ0@E2!,պ[䕞c*x4XK>BD{XMD\^G7U75, M<ʚlr8OEd:IdiX]D3Vs4ً>%Z%r!AiV&# b(BŻZw,ѫӦScuV}qWEb[u$ Lb9i%SL"C垁&Lqm]GSڠt싇2 [,}nȝg 0y >X*qn>+=ˁՑdYh8!lP-C=|* nGqy۪\u$ؒB7+=C*4r 9FB9AbBH^#K'`CB2.#8!䙋K}wZ2>6Bۛ0OdO>XZr§aZ2]Q9YA)~256tT lWQlG&Ml^T쳓y!4ޥ(uD(YO3(Q)g~1-#p™`D(L0rpu6*=q␼ PdU =0|?`6KV '92o'\e$ZY瀸EVGZTnU&"*2CgI=Z+&hwJ@0k֝{,` DŽa 9-)eG_nUO6 vP)ھGtnF_6Mm0W Y'b"ries))E1(ÉHqZ+Tl*>%"5’W.Ҋ u|:|?Dfoխ4t67o$Ue%lGwcn2+:\VUgO>Nw^msmQ6 uQ<~[sIh2O0BjF}L||@y;8.J2O2.?6HgP/%j.ip>kaˍ6Bꚕ.Qۧ27mm+׿FSm+wNuS +ʧ4m9֮mݟvZBuA^NK+z04E:a!mqwϿ8Y '>h?zHv qe[mE׬1y:w_jn8\`<v?e ;9o)wV#]VZ:5Mw\f=B-K]EZ Ե-;,/Fm~]t.&nlXCF a{N]oŊ\:QA;@ut#Jl7P; dL9is.iȈu,bI 4S^4wdTGO)fxSi +.=pD>teT,kocɢ*i!#x}Trܲ~ űެ&#I+~`ymo\ Y,p!"$ I9_p$ɑ4&m3_,.lz<7n#m8VRz׿s ~ }4D5<9y9ĕ'POJIdp,R&PK5q{_ezp*Ï+?ҺEW{l*~gz9Lf,Rq9.3\fRu;:l] /Njx WgW/֚=!s%!+TUV+qәgh0j ?J 8-b)A^9} '6L,GSG;أQT:쬤~;y67>}ox7K߼܊8砷xyo}ZIMM[s!a{5? ,>$E].ۍۋ_fs{DZqM~HgÅa.dr~7&L|8M{{ޒ\jy'W0儢eeMo6(Eor[X\'Mb+Kjf:nro.oJxaJ=A&*#_~l>+ :f}]ϭMSOqeGypc\%eX7i(¸?,fp_4J TBS D(|PPƤJ0\i$ev^ĸA{bIvMIwSkXvʩ9\sDI&!m\ia O$Qh8gQϽPdڢmu)uluN*@i]tSڈIB24$q2;+ƝNx!uPED%ҒJf76'vLfsl2z._-W FV#25uw IKtQyB?@FKѧqrKRtgqR;+iꏦHkRs40.@6Z:*1; o s0Ri(. ڻgs <7Vdu˙Jm XweD>Pi)U8渔q.]A$IJ6 H( 0cBbkô҄KnojSЎ]jF뷨6_k<oㄮjS{$Oz>˓FP>sOׅ01UĠi{ 0}i[a;ܕ8! ;ÏZqNy8t \!ۢVkAחB;@솎T Yf+~8)%% {a>:? ]HuQF7r!F Ge[M/FGJO; [݅MϿ5~Hλ6/q8%7:zTݛݞkkʻ«iepv[&^?Hv1Kj/O`>0x)YӒ"bmKBnb]3l}3Lmfyp;6j8+pp~Udf?9slnV3bIhC#aqu#S/G! g!UdaǪM\sS>/|5o鏮ڰ2C [2cv؋W*Y?L*7tUL笋r/G$'o_y6ɿɫO(3'W.WkE{`rn5oꦹ4hߦ]d״ 톋Xb7 g:z-gMa/ؽQ$\GNQbt%x!OkA ʠ,u.Dw:PԞm :JZ 6@ ΂ynM %V@!) 8AWS:=tyf:f؜x,v<=7:=jqVu;ScjK9Ziiw$ C)s Ɵ/S,~Y<\[+@D%ꓳk) Jxը6E"3|$uG7ztwޔ5˷< ܽ!-ݚ oǿQ6l:嵉V .)t tet(*Lp:W~]H}Rl;ҼAeIEcUR.K ?DK[,ZIQJжQhc6Rcxk 6P(謏hhdSV2Q0PlLbewLˡ,F h.禃.|K5ccem>Gm_03 b:HxǫxhO!/Χ T (c4Bp H|dLS+Lm,d{$K0edBʭїJ(M0<5mAZhFapJ&f&o^Pcp[*PHV'sd5yg+j6&N{6R73 %:xX""y˕1E';QnYGU߱q[H0 GWz/KP[pydήIIu9Mơ^2i~(/N% iI?}'%W Ne.3KX'~E0Ы%:z /fCsj2A03ɵ),]e~-!J S8MgjC\jɌ(_?C:ן(Pbk)J!DTeAF ~>Ddb?㞊.4rkNOalLw֝/r< kvxMm_4v]̓4;@uR5لk׷x|ެ&SvSr&kG/)!m-F&/Kʊn灟TAtD~*ww7?s0{,8k:jnM{ɋYe a<nx^}usdhm<Jf};<[nF7lOͺn7LϿ!Cng|r{Ө+Cmq.HHۚ3- |Eߛ`5Zq;)cca ,~뢦G,NSXG-ꣴ8g n(Bq#&OKD&\@BhKEh,;':5<ݩptϬGa2/$\YwcZ'CRh@Xz:yAS=҅$QII Cef]p#sUm;PWu3hs=‡~ᖪ3Logzg)=w}$Up}U"7yњ`.0)BV@Y]XmqA@4lcQS֕} -SF :y@J(#)E$@4Z5u(tzai` 4p|Rd5DFFs7JhxW-%xϜ N=Nhh IKeDq\ƢnH"Pi@1+>=WptM/"xBF"!R ]R2©]Oܵ9n^&j @jpLby*k1 ?""12K2Ddy+`?{H)w֫?wNyi5Q @1hP+rj@" ɒ&y:Bc-ұ1(z=AמuVm+֜mu0 >R8P ԍQbֵXO%,3_W>!D LCZ\)}]}a xi'3(2%L"x""xkY ^ܲZ !*'P+`%2`K+)B]\Ĺf&d =;/Tjٍ}_.kyQ"G/ bC:Q gI@&Ǚ9e>rXy41ͭ > 숀3*F/$yU&΃`WŌD[_ICHұYGP5LYB1F̛&2:'#T6AL2(gyz֘8[Y^0ѽox5.jIΜ"|چ⋎DѶjHWa4" Ŏl:8p/Uny¦8KV#\/'o$27N\*}?}e;i=MA*NQ/(Y~Kc@mo3>zGqpz;L2K|t a2 t(Y]3m#ϛ,&+ @5Ʃ#c,>Rs_TOVgM13='QǪ=p~>v`9r): GcRt&jN͇ IDubՂqu0!}L8ٓmHݼ5asu;eL˓|]p8ӎ8&sj[u[)>]k6,7slAd|7{`p @>V_T]~&/l/!6k%o ίW?Vu}7 G/YeIco%z)_.3jnt,I0Mq5P KVmGåtjTSTh.3P*8e%9Ϊ:HuN&Di~˥U快gY=˺ev=EɶgZĴLgZ0Df;${ jvE J%ўRj!% ҶJT9A ~ jb=k"y2QL1zu3'%GBKDɬ'!1#9R0&27uH-UI Sk k34tbܿ]@}"S޻\917t<ʤZQúewGO1b2)(N1O]SL5ډupt4/7Y~4 EFk"b\J?uO?~zg!Z+ǣxbuV[&h(JHB97E'㻔928E]-L%ʬbN +Kyw>_3E'n>~ϵ޲ vӯmn{jv{ʾ-,nn?d8a 3-RJєŀJN"!"KeçKeƣ-ZG]hF+3&O&0Z3&BR:|Z1q<x4Ue/V~lQWEƮ-g,;p9=`AkиA'uPiReAҐt(0>Frq7(DRo؟HɲRh7JdT$TADDn|A2+%34%@wKct1*SKKmE_%F/ Kҫ,0-& ʵ7t{7ٗ- !X)DiUd.O,ړN&c8cv @(a"=S)X4Yj@!d84X! JAxCL(}8}BqhP9;2u,& ޠJ̐ " EfB$4vs?6F}:>qo!!%q{D\TqES8MNYݖ 76CY EDk(!9==UQhҔ,e**y4!2$&NSaqFj8;ڧbdS\qh:X@LIJNN A8-A6Z \ܛXFǦx ! wa4r˟5Li3G޻"_dFz={?>$]y?fmwT^ё] ,jaӪFl P{]=CJw4k>DGкJխ>#.algg[U[T2>rW/PԴ\b5s)H˔,Lʈ[&#qxx9#hξZ2!:Bۛ0d(JY_'A2QY^1|S%tEv#:'yyMoټw֋[^w1ǝ&.FpF-)Q)a U[k4O/}ߥ{A~WgZ='ݸg7Io 1sy}|_ wͶH0TV^V[}Ni&+rċ,lvh74?wUny¦8KDz?J`>~|ĝv /In L,0=k?sp277 + &|8}yt2òeUWm6,{fy3#Xv=qСwVGZ?J>5;{fUx|<{/~>XaU.QVXv&jt>z6] .ϭbX/|Yi5D>G3AkhZ;B-H8[k?jJw!*Lrh0u>]|ImP>J(U>0/(M6w<%Fv#?o"x!m!9w)߆Gwg18+|g gYٷ&ʧҪ/f1]'tMųHȵ ZiR 'eA&%A:"="XC O Z's9FLT:o¥8k\pEQ`Z$1zԂ֜ITNΔt Sg!Lc ~=+"^`i); _W8?n|1YWKu55X=ËzvwʁKHbf\3F8!mC" j=2$&>PZh< =X ==6+|:xm3 8 F$Q2xHZy O. ?Cxⵞ%Jx)Gn&1Ĝڄ(&Qjkl\RA}ŴzaWzZه!z)*mٝĊ!d!Tijw> (|׾-AD}r(LzP^rVH(m7qk*{}`fpݒժ5bp:VdJXxH3TPt|~DpYvkmޭ^qȻ7ۂ3o QSR̲;%ʬf{]y19lϥuχ&'A+8FppǛsඍ]Tv5"7<8o8g{\|77q)j0ȟx'H_]v*=?Obu2mN'AÍikj٫ Q+'xk涇8vspϿ;s"m]tpݔH66>%> -_pKP"q["{{U]GJǨythtpO.gƄIOnw')W?e;lNGVu*i8.Kϟ7y7u)ӊ71Imr/f|_wCC GO&ߚ^{;-?9X-LG3L-ޗV󓗐z,K=-~OXG5u`J-Cc1'd6*fy)B~'|ﺌ}>w645 StMqg*|*3\ͽC~6ik'{&ʿ[\@U!3hۼU.&l& ٵkn Ty.;͚&ޭUW~r~;**-6j]+6+G0#D`*pl<mz2>:j TGGL3Ok\M{C&mȥ x r+D@v)1.e2AP&zh@pDgI3B/qyƥ{] )},.g94Y" N_uhw_>*U#gxqNhg0!1l'"gO>o)͸-'Jv vi-8Ol#hXdW Oݽ 27]1)0;+n.hUV\wCJWZT; >M~~[p;*gB}<3x=/xR=x`@EzQL=qȕXUv{*Rz( 9:tXUCWJ֢(ƌ:t ĎE\%jD%mKWLSDG#VԱD-a.+9W\N=;b64\J%w^_r8lvM{*9wP87²a4*}Ayՠ>k3۫*: &5UގLq`0ڸb<~1f Y僛_ﶼZx0~{K5Jk!zBq3 ָ h[:|rpPRRАИ3I}A08tIJB03٬2b{1`H0VE–=&v*0h5{J&T_D&}ذ(ƻVR|:Ozl"tE!`J"7Hp6o53fnk7VrolǾ_P"Fb??[3鮚3rL'oʅ:9F:Dni_ b.XS0$pQyڐ/1xJI9S:*V$n}9S2:QrEf GHJ'\F3jAĀQ (W&y FҎY | \KKȳU6_ir8Rʂ7#ʵ.!Z@eGsδ˭8< b-'KR Q-H$LGHM{b|ჩmα]ߘ,#DJNcRM&C93,*͸Jb,aG!{f#AX2T*a֗}}u$pn ,DDY XiPAˣjux):HAz50E#J6|,1Q!l8D,@6nf7.nXh&D I/^ O1HaW gac\'.w'o_ NoO- "UB?3eW?11Iγ~H0j(0 ITvS!%C1{PSx4w 5iJG;@ryB#dI@H6dD0!:RL݉)гۓb Bw:U85(R\oaA:J\$>\OW燁I'y 9tFU}]VWJqox:QNטt}jC&^=X1Soସ#$&Lg sZg6~-T?O<8O a\ֹg>u۹-ճ3I #Uz!!b0>jA)`*| ]:u'GQIu\)A E Is`o2M+fl0\)F39VJ[Jϳo>o:K״u\SyunftN:}G2X7M.K UNe2Nu:+$wo]ị7_`.~wo, qpp\$A?E |CO614WMfFg]-6&boˤz! 0n@-[#-Id/IGHE$iDmtTP(;1+ɭ1Z|HFQmOqZ \/*4G8pF"P5>Hǡ^{JOow>_IO,ªY< Fɽ%Xx)K?}6-rFBB An)19LH)hǤ `O*=ud8B1bua߽ɐ.ׄ &DyZѶ$DL ;가pA5=u IA? R3ظ,9ÞEL)lрk1F,aD`=[/ai5"kӅzha9-՚5oKQJ6ߔ"3ܹ@IƙY}RMd=ӂOdeJaUF܈>JձUewa]pqh9wZ=E8DKZ B Q+-0G'M$р8ZXs%isn(h SeqiUڰ%{AýuH ÌМ  +|H@qǫѩ; scz t.+: ܓoS4`bQ~L^ B5y@!MnOgGyP.]4w!4ÛQT,* `pUM L7`SE`'|SytY~-!9?B2ip•O 5M()&Ou3I1$SSQ?r$%B(%[u'luzw$Pf0*̒47N ']oA?(=W;9:(j6a-: ^+]Wg"kˮ6%ۚȵI/]IjS [x2W~xhCru>i{X~i֮fa]]z5~d`ٜy۹ OI{gkYOnmr6Oir8ɭ P7k"mZڞFftoO/WL^'[U Ƙcx4&q^r# s-9˝R 3=4۔)wժրbR~_;ʺw:ΡN5Y<UfBkº5*Cz Y@SJx->c]iziD2#Ԧym~hUS]_ t |} RNG>q6'Y4,r)ٻW?y7z}Z9\vtO7HU9m}du˝"2 Nvs肋ڣT8`_^qkDR`j:Rjsfȵ1O&zGkkm}!8j)#M:^XZբ(k=۵rw#]m i@ͻ~1eH u)4?T<*i11g{TRT{q{F4jD3sM 8'1`l,Bʄ~ -<ԀZBR CY+m#IsW;@czb=^vDdd˔ odaQEbSb1XqD9fC( 9D<U+, J]gq}g7qv욺@(J/c9ml4B\[ yiz'!7+B(Śc6$k1kER*ϥZp3&fCΠZokdƪS1+%XA!pM-V qKom7^}MweS,SmgLer jG3.zHG!jIJBPSmҘs >@e-MMYF#+QgوUBN.h[0+'FS]\zodñQ~T`Y؛"|( $=֜ejԦH0ԛBC |LXtկABIL6XRXI%"CZ ؓw5ۮu|nc}9wѲ@Tb,6R]Z"hw5yQG-r2o{Twi-lk3y`|p`#f\;[[ZL0,= ՠ}OvO~w~`Gֿ'Ϗا'Ţ>Nܨy?ҦGIiԥΞsQH־ev6.D~-{fI|*&o%VF[XaF9vN~FF&on$9vYI$ of?E:߬|>:;g/uS~-wn.c֊2zaǙhGZ6Ipyn/GrBy 9~uۗ ee€~+p+&͙%I&ǀeWU!J5<&ՍjOx4?:y߬O|,RvN>s, DxY|&W2sqptz~qHcĵnmjDZHۛ}8=d̾ BW) xeZ[SƂOhk7'iE@)BK **jH@t$;ol*}l#AȾ;:ch@ 2{`mJҡ`0ʐXNZߟ*L!cS`4d_#|A6u\=>^JNS6s~t#hr넄)7yrnOS*{^vS*JF78 JdufgS*J7t13Dy/YT!@բZX *[fdV2:qPkbdSU_MéDJܶ')[Emm)Z1e]RXED0) M' F6w5zDRuND"6)z!8 DBNwre{߻ r۟ 536UdTɌQghI(.Bi3XXo1{ekÛJ@E O^!)@PkCiqqp`6ξK Lg kQ7|–4s{cl𐡕9  ˯j35G̸+pW(q/zTtѿp+V*%|,19zTh ŕw\q^!Bl3s? H:9ibܧέ{h/фz޻8NؑxXc9z\x(aCl?v`E\ׅZ%,d&AѱPYO{({hξ^<Șkʘ/\AX&. *WmBŏbEC.>̋{$Cž5KtiT1׊xrdm40B ,-΄8;샡>&}~9#v"PmQ*5N1b8[C5RZ}=;b6*A,s^WT]C1h逞oT*=^Mv1zuѲa\:nzg*+{d|h4h{E'Z[\:+5_lڷ__Ǒ>0Tޡ1IR Xf ޅU,8-\ (ٮRd"FTUDl y A V :X;XSz#c7qvuЍtnq(X&,|R,H~Ec[ʂǁY'S}>+7y'οq6)d'N&Wɲ#h ,Dк `wbyYHWĶI$М= , Q4"oRfru9pJL8#6銋y(M;ESMfHn5h1)jmkPAКi(}}P0I,P;񢋭Q*T#7="NuлR8Էtc/DoꉈM8!ⶦ%& :r1pjY;TL :djy/"椕j]t^̭"Eg q(C mO*[h!jW ɹ!9]Ggʸ)MKE.ޙKElb Uq['G1K>-ɟ.>.LB/Mg<4@frʖqܩopqfGpK2b}! n~|GqczY':=btMY b[ 6Y))&΍=)DbN>]G27ݔFe8] f{<'|e|8Bp0gWيMXHw&THD*XI~ɬC[mg9>ss]n_^R(u{1h]BUQ6!Xl6r+sYa.ߎ$/\\>^OfEh'\5OV.Kq]2tʃG8>(\]KvzWg3ӈXgm|ӻ:ym$k1lvrG [n{gO-&ۗ0fm߾&]8'SwǖEv|vOt|ިhRU_ݶn~zDJwz'Õ.{(,JԁeP!rtOQJu*.K C՛ͽ~.} [dMB:`14Hr,Q)BKqmQ9f's?8!ڳnmG%mu™x^_ oryO+f1`RUC19l`0 EMV^i7G0O .ܱa 1G+Sj=k!Ie)pb/P :驅ʊ4sk3h&x0bw#j*kn h`&HxH3[J[H? /vT}]nKaqeoH/G?n=úzRL*n`\yk(;v:0!&8&<'w43!Ϸ<6a U *c !kv_כ88/5lpuo7BDJ#I9y33gJP>-%+[ 4sa ҢͥmQX3`}UVE VxkV]Rr)wkTg$?DY.g%UB .'ßu0r`a!m{?>V5LUP*EדM_|/b6]۷^OF Dx($S,26ˬbru/|xr˯K+}bֿpgs|[lf={'4_wݕFW2W[}<~MHyQ/|sl2kh~^|L޼Uҋ ~uI.R _tֿf2*XW&M-_*o^T>Z(lgy8A sJX[yɆچk;Û0ߜ"Y2 '^j~zt^ayæG~jf~'Y֢ Dx0qV^n޵J.`xz+Gt`E`[-Ql4NҕL/拆TlڭgC7{*omz#8|Z% hl;MTby./m^̙ 6mrQr޼.RRbFc62J\Dd1I\b$| %^6#mgRn'@<>.²Ƹf"y,H&qsI*.RiYeUx蘡t&Czhq1<ЊIf۞[r~sۇp}|RH밭3Ӌ5@3>855/Ŋ,\p*>[6yz>Ydb< &z /^L_^ =p箖fa~7L)QK8)ro۳ryP>ȗ(3R3!|nMIv9yZvr-=|w LP97^ŖrˏFy|VZ},wƀkh$>< xUT,+  m>VҬeI/k .Nj 4(܁8{kV;'BxCWWS ]!ZnNWRvtxJT GfUt >]5CiZFW]ɎVJ[ݙK>$)L9̓%y%ߥĖ } 'x`ZT &Diq )*9 xA2,^nz0(_FFb6^Hnҁb ƣt.}gA|uF)?pЉ_]B]iR*0I.IفsI} ^ޒ@TVx|J %hr#C`y0ou_d zd(߾zGJhsYZd!>d9BYxjvJ PDEaR Nj)djpr'%k#;%ZIYPJ֣ 3!4fְfv)fS* K++/thj;]JUIvtu:t)5{DW2 ]!\Mzڝ8e+A1# "\ ]!ZNWtJR0>]IE{CWRm{+Dy QZ ҕbX])8ǻBƛT;匴e%}Еf]!`wpy3husWt)ҕaү"sI#\K}+@kisWwtutea*U5te% ]W=UhztjrudjWU3HuWPZ.R JutUTu   (  mZ@n)c t>2GVB##!XH~L> [UuNvXZ(2KA)Y(bbHF6̘dQ,XS=cS.)yo @L@w^ zkmGt5/th ߝ%+.#VCWWK_ JU. J#BzCW/)gD::E(27tp7դt(iGWHWJR}J`E7t5t(JC"}ǻR ]!Z&NWRvtut2<+|(orW֊|۶SGWCW*u)(Pbގ$'?r&71x>8TR+(T< c"LؑԐH;,h"} FGJ^gƳA=_S<аgNS P3M\*"heh%6MQՋ~;:LEt]:IGH Ʌy(o~ϻ?N,86Y0_h88 ^}B^aFQPaz),.)ճ:# 1\k 80ˊֶf{H-L^~p x[敃?? !Β 'Ṱ}Bl#?p0 #9xMgЛld5kms-VGiv) {ҊVXՎR ~[L+bYMjrG[DhgOa|>^ ѿs.Y~[4E>{ o=;\ٖzVu\ͨ&ҬFԚ+&g"޶Vn[kN\8;#՘#u64v8>ov?J׍`JZoߣF|GDYEM μozz*o VmQZ{r9Jk8V7C BG% @:$4yRĽp3nDwW < pCKVwᴾnxzUެ(΃T]GAKB^UX]kW^EwtO'n\nȪYW!{O!{TP|JHU(7un vJI(gvY+Y}t|ڃϋLs1?XGFuFP92bO#C"\;2ښ\#cN]?26_ Z{6HtU`2ml8r'HY6m@n%ېnc6q# !Ji`ݥVV;\nqk"n},aj{uMs)Weiy[uִrZ.eɷA4S_) -UW.CfK/0ga=.;\gʹJt$H%څ@djV!'&L œ߂tkMp-%Fʶ(.F9OOJjo  QV6[tU5DWVT+atmv [ ]\ühw(JkO >{+ˬ/thOWRuO0q}.&De6AJ[O+l?v˽+DEDYy=Е|*d@2 \+|+@+i Qλz׸-w5[kVHv;,Qڕq` ݵ7|)De`J7^=vQ3@*8/tĮt!+DLGW'HW#po 2 ]!ZiNWNW\zDWo rXoȚd+I * "\!}+DZ Q )ҕbTRX>b}VT+0=|' _ Ѫ] ][tߕCL/BBWX'*6Birɓ+ˉ >Xo j ]!Z+NWR.xT.=~!C#{gfp}PI3HuWPmlЕjkՃ}r(LTsp o\@niFZsx@3 *iH.FHα5p1Sөse)YQ)ٗcO-XT(|fyѦ%6\!ʶeAb3ƭl-+O* J ]!ZNWR L2]!`wpm+D)ҕRH]`y51J ]!Zz P*j::A2*<+l7tp9rvBttut$ZyDW!6t(%RHWZJ| BJBWֶ Tvtute^VBBW޻B OLfAQ RvBLwt8a|P=!8i*pM4}z ibfr++>M z̜]cG9T1Jզz3DUWm;JZʪV'u`/o iUoŗ/Uu9peEܶ}E5l{:-SfH [ ]\M+DT Qr \]`)7tp7thv1Jk::Aib<+7tpQߐ5F{ÅSJa.FF$`&X9 (102g9-ONO鷈&dK*0H?71ИUds#Q6{ oT#FSHQ"^WJ*$~$MV1\dQ T,J&Kqnw|Bu!wys+ͳzΛk#iH%5Væ tF=‡zsNqN>I@{u38:Akc0mƀ6s VRyBHQͺ}!XKݷn(m|i"%VOwڳVj!G]чΨ|W!'c#ZZal%EtgMJb=w|֑]Zy;aF|Y[(P!6Ɂ)BAnQx4{4:<4Gda0P.A!V߸+A[,X]pl5Mm`1גuV6`Ņu ixu Ό,ecOD@ڶUYl(E;̡m*")roPP.*M Ü` fWH;i\2gR!}EБV6>N9G1jTPTߕziZ26HQc&#lmv^Eo{jPB]ɒ܁E7 +Z9q2FmͷA( `-ePBdBES 4 U.1a RLAn9 N0t掊CC\Lu݅Z'8TR 3r52D,R r 6lC@WJokLEwR )ti,BpSH( :j8ui0Օ/H!ZD5=( EOUbKRt ueFjuD$J(k! e 1P =(aL! #A ԙJzkt7h7X+f%$' cEUzP( /QyMgⅩ\]\5]gܨa} L0B6#&f3x:pvPQJ7#Jh+fFVW 2r i(yG a;X (ܬ`QBąAo%CJ$JE2e(xGZ1w g yy{:%+&CFPr#VH܁m CK.̪d!:T?Qy'*L l;VC]֫C(NCokd?XBߙwyv)Ru+yY}%BE4=R\mob:EjC$*jPKB*q`\GQ적PA,=kq[S.fX;6@D]CbA J5 cͪ)#F^ZFq/D(DxQ:Wnv63+I@̴촪J)C2~ȃ 8vGyQ9OEa2,Ƣ cA8gũBPˉuBR?WHgYtggMU@ѠTf5փPڀJ]knh[oMu^I"[BZ&7k! IHE}>F9K`o Q84e_ W7݋3S,|OmڋӾ\ϕI&~T`0u F^,4zV6X `[ U-4QG[s5|5ǤͨyHY#h4v˨ 4$/%зUfMj:j7LJȀ=aT9 ]Qn$PD;DTP=`P낂a KHπOts7=` U߼Yab6 V$k]ur7po-w$YycaUè˅PeQEҺb$U7{gus[sO*6>q?mAiumlb~v]ݬY u(P5jI:;1jGbj='е-9=Fw *6LLP3xFp-4p^ kB DoO i ֚&-;PZP^k䋑Q+Bj}@S AoQ%#p45<B%aPJ↑ttk ;"@.* #Yc6\W˥11DL0r";@joBER!6o.Xvk5c'ǢU~FzwW,"k0BI>z{& 7[x#7n/n9; x.hWͩN.P :uܩ(N_ZLWvrH\R @-G6~KޛK9Vg^Cn6i|r;O&==y[?<||3Fo5V=ĕ^]^8zB\} _5]]~_~CVJG_Zo\Ȅm7Gd헫1ڸnj@g@]ˤdw:}H'$'h9N n1N}K38Q:'iw|qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 D` b@ ۥ8;2*y&A:lr $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@:\'B+%9fYpb h(}'!:ۊH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 tN ^ !b@@;8 ;q $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@: 'л[[aSjʭ^o7lhw]\_ճG@"r]q)r]RT(bc\Z޸ĸtƥ & bD+Rwb ] ]A ~-GWB\ ]1ܴڨ^]1Jg}c{bBW6lR)QWHW)it;Cwrhޫ+FН?]~}v?n rX?Px0bhս;! - ]Y_]bO}jNh5FҕskCW 톖̾ 9‚؋F+R/eC+|K*8-h:b'6]1dCWbZt(ICHF Q]EW /fFt(m:JD!GCW w9mNW29? ]Ѫ}}@ J#g-)y>ToMћ?Lt7Yz' 'e>U Jڳ%;K[a̽GvjMvӳ[c@7Ʋt[@*3hiAI4!8 qoZݯ 1)qqu<3uq*coe}u}/˗/.3`dxoм_}}o]l7_׳ξ6/|s m[˻oxۀ[V:-7GdƯr^lbnh5OAݒ}~l=m><Sxͷas~y}jg{&>!;RRmQfnKju14a[tIiO6%ԣmڳ_6p6cn}}|kce%$~XB+pAJ^G|[=gxL0=]SU7<^FkZd1{4I+}>l4 KOid \>b<VWjԢ[ƦſbA׊Ja.h l@QU 5駓$Jݶ9$]xcFO h0/qQ+6UbV OvqtX;ZQkHKb }}~$Af*WoD҇Z_YeRb?,-?WwΦټCً1،* M> ~3QB=L>e%mM۔8?;X'}R+M .hx>CL<ВCh |NGhvfx^/ARv b]z&Y u<;FXtFC36 鼩N? 6~nVry_[vr0ŇEp0/Ws͓3P @\^%f*#ZsMXA&GLE1l{p+!r{ŰW hɹ#-A(!QΤR; hѫbE*L<-SDcӛgIY,-,xGI4.sc9W+v4! o JHO @jQM>:\]Iw@$iJ vg| wlC+4B#qBƔ+odM%gam0> M, GLfcgy~4CX#vF S=%"jKANV(%k^4Oƍq.){y'P3m[8%QpAb<`VʼN D^{s^;uZtuG!.PԷe7W=>x& /&SlYFt4zT0FZ-bDsM U񮀵;yNQqb֨s1r{V=7c5ʸ3ÇdUķoy1h=Ϸױٳ/N+cҠCb:{i8}p2%on>%LjN0Z]B1Xat Le[-ZK4xWB!V feںP_ Kgv1*amO5i]o4on)gOb*$؍nVG.#,/Qu\:茠d2(-rEQɯo֚bۋ.΀O|w=Iu|]2Mi&y|$Ԇm;xE۪}uPm\\ji-ǹ6LP]j^7^޺qMTQ6 9Q(;,n׉r#$"BkV8xy*_]70Xae 8!;mTcrϔQP:8b`{N/nb/r=ʧHk{uX49B`"0^vUZI3nn;YO+Db:"!9X:^ gL ]K |uVQ)n7G"ͧn¼:JfβkE\:< b-'KӎV&p(`0$LGHdbp!u.KװTxP")9IE6[fEJb,A"A/ڄ WKơ:4h7wTaVBT+IF;2HDU+ !`yT"Xm߳:kw4]GHk`=G `5"1i1Q!lXD,@5~iwX(5QH"WJLa/+Ƴ01N׿K$_?ܔv0 .,= ̫Iede⯯\ DbB1̳2/;f4IX~Rl K^HoӔ#"! c? Ai$OdX12`)"StQ J1u(‰sP'a85R\^KrLg <y91itפC?ljY}|+5`p8?/Jq9K@kՂ.)CKG(rhǰم1յzrM*y^<^̏ O21qy>'pGÏR}>- b^N،; G b=1nHw7Dtvfw!;2:jRmVxtLtqd848=86 ׮R_Z$vhd$ú}~3IӶ-%O!>٤ryks=V[jYJ?d_aRYlQ \0LKQb|sKLu\eZQ醡 ߫!UqcٓϟG;D=ѻa-t~BwR2ZGi2%ao|z?UL ߏT9ʣsS[D!\zN&ɭzF7iCD@7#W;7/ތZIo՛QZa6,EDK.Rr\!g2z ip!-hAL lUWCqV>&-][;KX 06 ߽ɐbׄ &DyZqmQ?NMv+N-|=Dl9T{0vw^S& +SФf ,9"h01|Yb# :zl'E4XXVb臋?MVlmo fbQ?-B7(tL- cu,k7aO~[B(Y{a}|SpXOC;n%ޟ4֊qs;:A(w`i"ϿtvsS\qB~]LU"[]]T~UHW}ym+Qm ޲=NC q&ܹa|껐9[SOlY;]R55K^biu@D+/)w}~q滃Gnjxx|V{y  뫯8j{|2]Ky{;HOtwPx=v?3g^(fc0&.t"sl9fL[dSlP8xʃjƼ|^f;& SI,*FJ"!FE\3rX8g[/^04:W3٘lm1@.. { =&d ;_N[ݺG*@wDqg|D}R*W&rl*@){" ]r&2oM6+f?~TJA14;q+y\s@ ul>|2E+BF:ZgC6?V!a_yFfj|BZ5:C agK,fˁz t[ E`l$A3֨c^=)o#qv.2oqsꕓr}I.4DIxc` 2;"< Pk*g3OywLҟeJC>R, -jm5Dx id]ЗΣ9p zƲ^xSZUac*SeBIO.͕3߇>48S֑Olj7'MR) K{RiR(vt!ϝQ>^թ86Du>?nIdx%o Eo 6Fm)Ppz+iNt5[w !zGgLvUBgeiyZV_ a.6yb6>޳՛m *E qB!7^ye}Ϫ`6dOQbZCEb5Jkdk[35F5\q2UM⥚eG3n:SڟZ_^ORvFbshŔGD:Rۢ' bu7]RE4%+d1HKƄJ9OJ(Y% 4Zx9_ Ǻgw{tυ۪>YSrهTGIV+AaP5J9eCoh25,$SDEI,S/zΒS R|qF=u%I)ԍxQ~`wa)DzfN̽Pbcۇn'E1m6O$MV14~h^Fvn!{T!?MٹoG`Z)%a#+1@+SI^V*0e!)b%eN B|VjdU%$qR(җB ăA0svZcM_[NYu| ;;?H|Іu9 vK]㌎^6t$\o5@ynrל blHEyj I'JJVSH~I-n!A'Nߝ|<_]7דVcũo窷'mt my QmEr.!⦄zQwNg \OzɧwZ' R_NN~[o2wg?=ճd ®nmjpmw/@?3jPᶋ~p{AcAʧRwIZ=X4N=9ΐ UDxE `^hB77kM&8vфnevhO(^od;vqs/W5 mczQB?m%znbq;7lߧhȢM)lI bI%\En,[(BGzʡ1DIeJX =jaQ'-ZCUGcmN02,  5-v#--8Wh۽qqyQn*|7]|:;]\ۦQI&YJt!+4FĽXL:OЈZ=kuMb %,)3)7bfJL#0svGl8{Xqh΃v`ph$ޔ'\D x|T8K)CVd,5H5%j7[Q4/(Fwfx$5թ&"E1##|sD#h6Q0%BbutgS%+g?T\FĜz勹uq#h[3,\Pt*u-(et#a]GӍsSs(94.s\_bɾ(A6PM>{g b>d#dCїX3SgK9.>G\<@Fǡx !l]E~XΑ{ 27bZn\Q,(_78Uxlnkӿ??=i5Ȋ4a"?%aZeQx}'o =޷M(]iS..QߗۤcѲuɨamNTj C,\$&_Ig[ݭ)6tXC}cn_ Q(z,Q1%'gz8W! ֩:"@CObQ}SB:A{RXSBW]:U]\@!#sH=ĐLL8X}qi4ΕGЗٞCsHOڵ s.04 fvp\Vv0Z =r\ugA55+g9Xr[ѕ&]WNyueTАqShEWN\yu%JHڐ q5+jוS"ZTE+=v]94+PNV]-QW9lHW,q3+#mOוS⺉ף+<< ~НǙ3_~\>7']}$to><ij@<41~sv}u{U{h EsJǗ[vBחۂmݔCwg}y{NW?-@ oYw+q\[3|7J# GC3Vﶏo~yۻn5ߜ;~X}g|՟}'8myoH#v2LJoT,`>c.iyr'd()@]Y2p2[%rlHW 4+=y)׮+jJֻȩ!]0PlFWChEWNSrʝYR+ њter3rLv )uͮ+ $ҕv+@iϮEWT94+AѕԊv]9%%JH[ 0S;r\m3h?tJ\ .QW Iҕ'ft帩ikוS\2(JK R;ѕr~)#jtEV=WDSO~WWhӉQRKW4CWU,k]7 g`"sȑޥ(3J-́RCvJ $3-^^CqWc7dˆrq<}|o!~Fa00u]A~![?:3:kR*9Bwfc` & WS*YvsH%ʯ5:FSXmd2Lr.GDy|5KC=39nVzf)jyמ{f1*17+fftLitt]Hurtҳ #OWNg<\h\OוSF\u@]a!]90+ftCu唵*Xu,[r`m3踙[ѕ u唉W]-PWy %4+ҊVbcWNNZ ΠMtZ3蔒V]-PWҕ]ѕjCFiXu@]eZ2D،W -BٕS⪫GW|`y@azg쬉/0dLwǏbStexr-hUW]9.7+\RJ?Lӕ?cqc3rZ )k;NjճJSN" ʀGԕFiEWNՏ]9%cWKԕ=SnFWڊVv]9eUW_~u5 8é7H{j-ly\PЕz| 'e_Hhs o ߌr Jhl &;s P:eNJݎ7ml[7W?߾&pJ?2tHnc)e "TMZ))"=66bվ_Ç|c_yu^_]^nWg^WvuOߧ^?7Wtnp݆o{^`n>."ldiho(C I5>ȝ՗|C1\LV\e=uڶ^gIp#fnHW܌Wr+r̵(ª*QCNҎ ZѕӞj.2ݕ,+nɠ*+PؕS"]>XHgQe@3-;CgV DӍ NzSO\C8m:N4pe,3м6WwWΖBs' -{r5fw֚m]HL؀LqUZɄVN<ԭ %EM RCr`qOo-u唵\u,J9@CroX8W]%v]9:θH]j3hR3r\lFWN+PReuEwhHWqKW{곁jA̫+Ymi҉Cft،SrJUW ԕd†te :njfRcWN)uC ʀU]nf+:e^FWz! LSE<WOAS-+>u>|Ƿa^|gNy]iOkjDx?*{aaY52 v_H>Pu˲>3pL{#ͧsKu߶XAwֲ⫤>)6r3n?\6?yn^H|DIR$F`MC!be3M~*w|7g/H~7oq?WuƊauuyv ǩ@ '2:uuIrQ/u@ r e%;ס!;PBB)b?8%{8I ;}(&+$A~\hq 0ט![)Ѯ7Se H[ {i1+Ǥq,h5hqZJ[ed@!Z1,i2u%c"掠K)szǷR틥i".&@D!2)3e aI d^Xd7U#Eڤ4t!`}5O}WJ`{O@LD3fݚqSd:.begM? chEG%BMVUf`w!E j׏VvWQ*4L}v8yxȒ٢D,O@U!#\0޾,"-=Ğtd;,B]1ߢ|"ң9ͽwUnBC8 hMI_`S&fܔ+1@BO@U{"d= 4- `S qk֭1ڝUԇQLDŽ f5kR M1 K=v]J40Ē! b>)ABocAJI2%rRN˰6XbWLњhch/#X|ae`v;aFX{O-R*!g4S%X!ZG`| r-Orē}k܎Ӹ?PCZ&q ,V;RN&+KK%'q(voM!e}v#nS֣3.hwH Olu=i ]8h^NlI񽰁,[k8 S_n`Yt-zd8XEJA`J}{MY嘰]'* )5ZH{#Øf̽uFk6`V{%B>&| փnPkY_}gY&`)ߤ"cd/,$Rq342!Kr"{`@k-U]:`xغ)!@1a%ހEeԉ4O<ĈCgj?]ńg)^Xg#|.;AwmY~i"a1M0W[ZӤ#Rv< =(Q-vtcĉf[[ݧ1S|bDJpPgSO nrP`=2t/Q*ACmj((Sѝ E4GRF U BH( {/u6d^i\F*^eFTU]%їTh>#i̼B)U3_Mr^KYvJDA !&dYh$kBҰ*4]AՊX>sAv ԙy _.ѭnz1#.EVz44U ULl^vNR'D̿h.`V9m:ter9_5ޘj*赫3[ݗe&}6}`-ڹC -A7 /ፍPMJ)h+>$ B+X(ITxmBV5 e4/! OݗUd 2inuG-=2p U?jPJ4%wV4`Ye>ZbM=A!%DB>hwAjyq:EjC$*KtPK|̡P}ֽQ@R@"Pc*2;"<6CܖSBEk .IҎa<P9@1^!B !=f br߲=V> (DhQ:fƮn6 sPHQ-fWbd@BUq@ơ"2ΪU%CʰE(3A68F)ز(jS,Ԛ}-6?vXI+Ytg#MGhTf%Ei*5^ZVE*~'XHhZ6H*tM>+%L0A7lmRjh\ cs^y63.IM;?MۆsU&6U ԭGnJ''f= ]Z`w- Nd4QGmC^k Q$I"eCkBmVh#@Oz~{4l}˃ʌ' <%2`ж+9]PnDV]qnQD{}I,WR SAA ,3 R4Č,mAz |"2PʽwƮXBZyE!8ij)K:)Hvr;X._* Q 2ݡPʢ#6:XL&ݝA,+~x$mDVc-S{N낙٭EuЬUP%HI21KCR 9Ok{i<5^h5э7[A[_%pg^qE D+ (|<U8P(-lZj>W&Eiq#f0YJFDyStE8 8%=kCWQ[:^#@.* w6\4*f\"AC, c4&="vSJ,}褀%a4QAygjp N}TQ/zi?~7fգ Z;KgLFMð*a۟~~OwP]rBwC^j(x֮nWz60]IvY) !&zCgٛѪI'W/9ڦǓ`@zi18bn>>=ʟ|J+=V8\|'Ŝδ̷O]_8>nEo|ts2q| m[c[ާ9ڸlt1Z>x#ԣVSc%FWrKpzq @ d!"; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@ (C gMp )~ 3B>O'Pb'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v=g'PCrBph8z'PRHv=G'Pt{v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b';dN yOW< `#;:Ck8C $pIv{+f';H>е +0@.p ]N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; tN y8'З?vX~iJR׷ VGev9]< $KH6 Ǹ> ƸD.+s%TK[j3[)u\ mݯkx!֡MxV6YfU 6fFŽ =wF?/bu(ZEmNNSc nQo}8,*>iͳߗs݌R?^N dZ?J~Z] O[2jeOxL P|\AJ7)gp+#<c&Pf!He4s~Z'/'6$w/[R Z8HW+~\l9GYOznqn?*L =)ecOܙ]S;gr?jd]r3ȱz+K]r] A%"YNKR9MkfLmFpJmFh}<ڌPn'pm|j3bTn@tEpFiBWVc+B3+̀Jkkp^Ep ]Z뎝sLWϐ Ր{1Z̍:BPB_1]=CFD-DW8Е5R3"A?LWϐ.I]WY" 񿿙PZ.#]y++` pp >+Bis@!``8"ǿv9t*Z]0ʫGu]J4t9<1akc;qXr'>h}a'fv;z?^y[ {[Iihz?M=hZ1M?x=TF&#}>+?Bng]٘{GLyu(<`6f7C4}g]9v7r} g.7r@ٖG1lm6HsٖP*ζ1*W+~kPІ/1]=C^DW, ]8Z4Z3]=C q8h`,pNWR+gHW;;5"~0FqtZ0]=Gr8 ``?"AB%3+VCRWp!Vƣƣ/ikWϑB2 i +PR;Pj9U ޺!-P1HpS Z]]yJs2n۠Ŏd8EQJ.3Pث]NxFe xJg{}C}X'lwEIql.<{iA=^Jn| ZBK'CGm6{ɻ,x^͇Nn'h:)@ ^v(gWr7m菄˧/fsliGoӃC-|oѺTͱWk/|}Cmb^{ˣ(7|9x5F~irVw<_^_ÀGdz9Y_g; r]oV~s5?o6'߻#zq^]\A|%im=o~\l۝⼛OOuҦXD)6 KԠʠu%SƊ[Liz^Ӽu?w{s":}󳤬7SW.M${7 WrՉPu-hC fO ƞe?ӴW<^ WaLߛίt~K#2?PM.߶ko[eZmY4F ;с7fˋ__rK6݂}8EO Ox-ގ\LBRFK7QE?L?~y)E#1-_W0(JOfX.Z)Ʉ㇖f7*}rujjDDz}u,W^"Zp%?#}4rs:nlyq{w<݊8ht\ŗвu1-F+eK6GI=g+B+JX*aSGX:{HϚLj$S,++uq(6>/FeZx?Xjnv6Ǯj)rx;j5K-.ޞOӕY~~?GQRwl#]ɭm-A"OsMfDuMG GRht%VRwjl~>LKvOM\]/D 3f[jՑ{Ë&I?Pˏ[Zj+ސSwRTO$?y3M:ϡn ͷ+.VYPev,>]b`k;ٲz"e}ܲ1Ji) lN>?}YNw IccҦ,nT//fmtfJ#!}nzO64)Z6kJ֢82ކdS6Yh5|UjHAomjkcjUf{#]ozg-Je !^ ݫ<? n01I.︭9io}0SwXuFY3KFzkq8ʝHѻ;sW3 Ruq/ߩ<+`D3H%%Tΐbճr}){ZU}mDԔTߤLYg*?=ϣ9Ȝtъ,uQ YMUu޵đ,鿂<5bi+h^^v:Ȉ9fm'`Ӏ4^tUWGWE|}ΰ4ù&z22~P?+_Zp9D ,*eOXWg[Ȥ0b`9̜!-t8 !:r,ޯk/ݕk"ZF8g5}kv{w%@v&Z.ǜKt,& rHx<4[QW֦Z{շ\cU-JiٹؒW"̹;oY/ gAA~:lVvAS,-bVO,WzF(`YF͍,>U{毲2Zy7YF錎Yqx"lI^bZ]pӅO?4!METR'U!b9i"v;MΥ4?)ҭ2W쒯9 ,%+)GlwO)h8ƒ"gt/Xkk^UujԶ\}mds)&ڦM ):*{Gsm:dO'LJ!ÈšݒߔSiD~0^`q֗5džCOHoQ&9VIՙPZ4}]F4h89&4kƊI|,.l5zs|ԂR[@(PP'bkVXf=_R9Rv'R0'Seou9sTNHC)J6ïx0!tC]]`EG?ks)5"A&S\ 1D;h:1^`sR$B1d/:Q˹dm7WLSm$>vL<؂{^*/:־ oĶ-ɠyif/wюaF=hX/ zôZRs}q}yN͉-)m.&0.+IP=NC_ qlK!Xn..$ {`5:d cPSVm \s7K.av:o\#q=gNCW fM2b}F1UWd xWj/TT%}#wXпl0D=yPU<5c s[VEb1F2աqF:+OX0%l4]usH9%"90sg˸.Yﭛ2QpRN@&o(e֬>شm$rC`d){SJ'$QHEEsBjhћej ZpINI-R!ΉMQ!":+C-OlTwqqa!qg']njM|Z6sSfonɲ!ۆiCgܯR}ic߈y='}k ȴ8hMi)^d 5ȡFb}|tWol|qpȈ*ޠJ$Jh,'IM K@"?̩qIƦv|qWclipt1 z~xptiyʴ:S#[$d]0sqhľ P8F EP}1&%T`GPPHBlPJ u0{!KK73v8Ӫ`65ESbQ.%Lm,,!se7S9(ӀUFD gKEJ#V%Pk1qQ}SI%ܘO 6qBV} = w[zfN=Qb.zxtéLMǸpI|NMfsMѿǦy4aw2[MsR؄t6c%-Cl5);*Uf=bPLGjXc<1D5BUAz`7̜{ [!{ge`އ]}>"Ȃc>mjoL#Fl1<'voQX&24P9\ m8'58Iv>RMIƔ68Psz7PZU1-'S52OeWY] Gg'Cgosٲc:=.J=}t|v/vƭTd 7gR @ -R1A.d(2HMkMJ!)MeIEGTDddkcЈum4231qaX83v ˌvvޙ/[B=]pv_0ÃӳخTd L^jeV&k@uWl(m#5c7-jކ.λ>@UʢΔ*.T5Ic1-2n#"1wyǶ]vQ{fzk#:YKQU3Yl&r29$ɶԗ[hcPbаhHBVfh9+KbH4՜aZa܍KQ "wGV_3/7#⌈&ÿm($-h!URXkU5C/`'>kckLbGՋUzK!I$l*(¹&_mhaܣ-Fi\%voʭƦ\X7}Ӱ϶ȱe`tθ35yc[-t8n_{g_]p){ʛDIw%&1϶.M~7O!4/¹V#Qgnl׾Ѯ/1{.B9Jgc?Z}H`0qKɓT.6LM Ȥ%ڠ?I;=~Jr$LєK rg7/cx11xA nn,{ëϟ^]K7}8s[q/>LJߎ5 0s_A&5U|3exrS ^rfΥd zqU_c/8:t,c7E}׋|y[~Xx=vɀtZb >}{]|OsWG~_~P{$}|_ۯNo' {{z.3?7}<>_\bo5ӫo J8ʸLJN#KT" -:wi(ze6S_k>=0s17$b=Ve[cmL/ls+=p 9FsJ$T+bix*d/_/6M]>4uS[49堂}xG{gz^։_?]%{?e/;:pa>ZK:wQnvp<&]{Rx;-`f?>];>f2rҥL] ~Uc(SC.q8;~4]a2G|Fiʇ!tNYӔ) agF=vҾV;?Sg8* yWPMK}M \^ęWIkVQ$KHeSlyU)fޛayZ1*ϼJ~s VAPK?fx^{VU?%ž.~U]lUjQӛ\ nd|Қ*}H#͐NTBSՄ8[NX2bfw.KҭꩇyP$slh!ڣw_p ̇lV+\^^[yo{q?𪘕,aYYd0+#hθ2*Zc<:?؍dj~Mqri&?O~#8O_,_ֻF>-WWُYq2/ nV۾fo/'w_4?|]'fXg? &üȇp9?w[s9^/姫a(N֖Lz"G?I*dŪWj,Y?;d@v}]Xf>WaI6Mt!Ź oqָO8ƌ"[J{N'WA(϶\eȿnx<c/ Qs+-_,q8{*+ϯ`{?u)80Z]5Q-讔v(6*sz:_ךlTGp2@u=AkL,NQ-U M\\QP7XVӬgE@47Bꑛq-'zr^ +ch @xZxkwQmG? qkv'Ҍ`HڥSPZ AYZ9B g )Gw]<ϋ0+7lIfӪG;!(%9D4ظ `/@yCnA %0[GC eݲLywPe*@t ὰlo oMmqߋO54[("Y)iD#b 6qJ!84˧^4Et5uً[ׯnRoJBHP G9 D D%fBQJ, iRFW1UbckbOfĺ}m=|Xʪ푘7sܥg)`qN`2K"7C0p/ _i@B6ƒE L!tE$K,ܧGIZYqc*bTJ8`Ls r3HtuH#vZer#9SZs`i=cB)7j<[֑aYh`VJHH!A~jZ٣TM^>{&[tgkQ5/67Ygb}4ъ*^M /&"}%Q:g\slGpzbHm9o|O,sD ţWj+1ߐB`*!e A*Wa x$ )[nZ D佖豉hj4BZ"ӄ{ V*)%0It2҈yj) q3B|0G?=a-}@U~攰4{U:vKr I U2U1=.n~@%[E8N2B_MΧƇ3< &7ѧ3ɣ ~lqNzXKn/&Bk(LcrK_CgSO|…U n㠨R$r)Y`p$fPt0aQ?JޣezlY:6t.6s_'~{(jw6DcLvL>.4xuInI Y;V\Z%;U|bXsmţ.=sOT:J7'ѼrخV{ުyY _M1wx|G}-3߷ ߿׶g]=׹1n v\znKsyѻ+~GUO~~ݱﰝJ.U=Li{.M+ ԙܯ ]*ܯR!~}_RJӧJVǷ'C3 d'qr74a!d 9Px!I[zuPۮ]#ZQq [Sg*0h5LuJ&DVm![䄎vb~Xu:.|2¤w8hqHҸ prbj`3l\q]U\u]/(K[+>w-yTmhNP|jů yM+ *#eΔN e5pmNk ^H"XFf GHJr I4rr{̨b qIr!`BL1jveZgAkYFq88Ɩخ(H`~J|tŒ; B[Zmm-]nklo>dZqIӅx4tlsp99s꼓ZWJV=|O]:",iJ1Zұ)4fPtq= Tꂓ zIK@V(-+uiPd6uPazGWԏUu $7|6ٿ0Qgo~speWG[!GCQ[M7MQSMC{b4hߥ]dvvݥ/Ib?F13v unAr9m|rvQk=(EIZ#,!6: P ϝ֘^NzjÜ&7S˃qe /uU #Q4A8#F *@qi^U8[ƹN#s" :[՛<D\O__~M$@рmP9 \-S.' ɞ*('{ꠜɂrrb.9U06 {镓!];  M"H牵H;ccܪlq _87Kss;qaMSetTj8rس)m4Z `؈%l_†y\E歹?ޝ7jjx(k\l)37囒d՝47/j"x"s.a Dr#J(aTw_j]Ѷlu{؁<ҼgPBG%J(I 4 Nc \ɨpleu4D{TrU'n̝vX-m=CJfD ^0Ś.QjtaJnhW^}Si o[y"'}lV'Vʯ30a!y`Ieg%3QQ&}Ժ&Xk!r%CO@_"W f]!WcvrPRړHRT:D`A;CNhj;J( $W )λ ;C\BZR=z9Jc$:% `I# &v\%>OJ!{r!WlǩgG%z`\= .K=a(ytWj@zUs6i~_/?ެ܄ c' OÛWty~,& r3񗪸OM_O7(.@ PFSp/g]U? aޅ_-=4ow]tǃ+Т}ݎf\ˋN =ǎѿb>Xip@ŔiUn${_Tx!Hnݩn^jh5DBm嫍5pY`ӄYתgsrzqRsw0)}7Y38gnVߕK9†W?:/ģ\Ԉj=o>nzؖœm9=i%8x;p%ZCZlQrFk #+{?\b+,(HrC2aW\sjX+*U]L #XxF.D "GWD)Lc %W,FX i%Q]QADrƳ%pE=\J[_\7& - Ӝq6o hUFF{!~2~.kCMǛhM9=qTӰvETNrkS1?xz[u5{[̥޶C䊀 FXhUʏ$W=+IM1EWJO(W+!"ZC+ gg+|U \)O2HB"WDkORtGfXKFWG QZHrC2YGטXh]Hx(Wɸ+*v mQ^8 D$W c6"\b+\SIz(Waxe"+VLF#Wˣ$Z Qjȕhz:+PpS #!8b8HǷZLx%Z Z+/..f;o>;9uk?3rc^&Y6TݻWPvv0=|~r3?JXnf=r3rǚ^0˥HX\2Z]W'\IέhpEV(C[dYJl*"B``,"}+GWD"$W"W[MDrE{?=F]3˕:\\=\es\fhe Q*䪇re"v6]&"ZRZ䪏r²A*+iy^(%KrCrq EDcWku,rE:؋Rt4=0)wzv> װ=W<&򣔁 rImzekex[;H#`0z>v[!-q/J0:'E@KwR?Fm v*(q* ZiB2Re0I#+VF#W+M,rEF.WD .UJ*.v\!~F"WDXrڐsg+Hi\uѺA4)\ie5+ˆ.W֪([\G u1ɕьE$W+'Z\eh[}\=\Y-IXx B,rEB+T)\V`bZ.nŅpy43D+$J\PB]gfpEv_Q(e|Ar:nn3wl\ngeX; {Nhdڏ2fC]GZ3b"ƄX˫Xc[]Z['֠ x,ޚ.)Mv>hxg+"u*=yQv|Tr+HQTynGp7b\kML9 #=\oBfa72gxcYx 0|O;uFs? 'Dxܼl~:x3g?N&ŧwu}f(6CS5~|Czݾy"L ٝ5s9+GBօTiyCQdZZIpŌ^p'[U:vFlmOd<̺wplD1^y,1 zzL0Rp\qт ]$W=+i,""`Pj(9$\)V]!00\.ױJ\2Iz(Wr""V0rъvIz(W*:"B`\4rE&dh!(MrChBE$Wl㉮ױX i Q4G,)$`P;](:^ʕ2v1mK"`\.X i(yz1r%X7 v%wg:jz9MG}ڍkBxmv|aT<-62<͞ĵ-GI3GXXG=X|6E1=/ :\!c6t"K&z\!0,"\ ([>?r%gBD$Wl]4rE\!-g2t"J>ʕrR+v:B}LI+yrE%\i]RhƮ/WD 2U8*Av]!`DWDMrE$\c "\+c+u,tBJxʕc똒Av,d@\j'ZB+)z9r;0'?`|'+/zҊaL \$W66H2?@se$@ۏoRƯ]5 JZpenK%iXc{9Xp˜Z9î:& EDhr3<܌(C[drgB䊀E#W+EV/WD UJr?Bةx+5*"Zp˕䚱4KjIH\!G )A%\)i+U:"ZlUHr#L+dzMzQ8\p|=/Ϛɀf ='F;Ob'LѨ.L(/ڢ; EiUV*'M^V3$MtWޠNѿ[ eUWyVNOմO/Bmn[%v.٦!F8ir^_TY[.3P6Lc;h׌yH׏iN2;ߍa _~zP.UvqnXQšL3k:4{\5@&}Xy{'f%`qs!Vմ2!vBq5]k.+ZM5=Zbw=ggl6~=(}>_`KEM󘠝GKCXcj:B }N&0ł'mZ'}+h|<5TP3/k4L9X1!5wr嚂ax $٥L/V~638Uf+1TȊKR 0LFj}ïl~es#Z@s\͝\wM| nC@̇EZz@{}7Ғ83/G@h#\{Hwwr5aJmʈ \brf^d}ih/UqEïf!;DNW͈DH\/^0GדD̿'y90|oWOyuvd6.!vYɺ̴L!}PL]%8SM?y> ; o /0^F?Q?2BBnl_O3왍#͚1^GSf"ˋ^N'4|Slr>Qgq7l]/&4pw>tX9[n& 7YƗ (6(b}$`mZ0eOl)-wir?iM̊ve)~j{xȭQ>TkVJk}5r0BdgJ.BÝ[ΉV)姱\ B -B'3rL?R[mNIŊi5i95+Dĵ,YȋYb pꈕbp'k.IcZmooQ̼\lf18yfrٓQ .}"N׬M6;נjUOtwa;`GT {yu%Jtt(q4~=oXgGq -av֠8i_cx.csuNȧ~0-}E.8p>g/]UΫiR;y|#t, FqMbxi١gGJՅ.;)?~,"}HacIu]rU\ՙ*FzZh1* :Y~㿽:Zn_ vx)e}oniA;oyR:VښJi \SQ?g & t:, Dm魑tNm^VBrnm e5*]^LP])^Dx(&M t$&^F62E6Ĭ^neYcRĬO9(ʘut_GV6Y]Gm\Zc`NLc\}S"=sǬ^m[,`*Y7Ѭa+ :a1fp_cwyI|Y;̧32/(d积_29Sݍr JH031MYO*J*-=ϭx׳rx0)Oϡ&+ĘRg?Ĕgs&c`!.>6Ps1 *$a/7q?k?@*M^Gѻ%7Àa fpnLΠeACLYɁ~G$B+ GXHGʏIrykEho<20^"r0QPbJ|Qg+8[ /koZPb<?$Z,;IV/ T#_cϣˆ7"2JFZ~w5MP\~yqy fl-|Vi.?e]K5c^ .B3dw_F2pF|ʢסlJM\ flJ\!gl9㬯Ȇh9 ^ ^lg Ԡf<6,?dOd")t-n1`Lb%[\A73Pi›JaыyZ  .dS3=hFc ӹQ*Sb!L΄dIB2'9D9%|rMӊ C>s3kpl FT{1]m_|FyK¤6,VI2K2__9[B: o//ʿ?\vaX\^&K}x˸^b=(YfrXgK E6($Lo_kY{%1/| XP_3.ɘo(fo;3`*hxwX9S?o?ߌr|3]'VJ;'781Y_G YkvX9~zSUeӏ >\&j0wgFa+7 }><]+=S^_Uސc991LL.3j0i#ތT5\mdFzGsBP|V/!MHgmı|o8>LVsA{vۯS k¾ F&MTq~WVd.2!^ڝifê,ev9-kL?/>&n}QLK@~X$|>_Lͭ!z˫E)Q&?2 PVS8³[=n v9"&ga%+wUK0+eYgZ,L8u VlcExT✈?S7³QO5Q5/C#<I_d`\>~FF%(dW(D!)!+\F.TFOBbѹn1lTn'l(5~⦾PX@)[Ap3_2;xsޟN_2kgЦd8R cЁaJ*9Heׂэ3n(F.9}DN/.Pr~m&#zCq)Si%7);g[!!"tp.jD T?( iL^ 7A  D g]P.D7Lj%6_ b,s|{l~kDtќA7Q~ܽa BP(!iuAoyjuklG?!*5]mvR!X ,sԵB[ ADg6_P ǐ ^]$Ifj)$ґj di Iq(C҇}[do[ 3+LvbRg+Ƴ6!(:n[XwhDRZ:-xx1F i!Ez$UPjVs 4Z7ns՘an/KuSpM zS&] fZ'$FDpC (MM2 ^81|>IrP,"n<ꤖ-9h٠5cGǷX5]΍'ۑri;Œs%> [[ 1A/F f$qϚS Nn㤢E:zq &.v{ݼSsEuT7qiUqT&0}b.`z0PYsdc"ǠD(14HEUGt8ņK7W1:0KxiGCwm{)"L4-yHE#|>E^oS1Иy81{x/-=NJo46J<%8${WRZRHH>=*)#n÷#șЕk@-'N܌! Jb6*Ά05A6Tz+/LrʹES_ <_({;Kbis)N30!ŞzR'LtBcDDFD )ƭYl9%6:h B 0B?w7'[ɿvJbn3ߐH=bD Zusn#0R V?8u[-|QDcnZjz;Z sJl?Bp4S:!$i]rdc:dLG5Y;gr1UAIx߶i1umX+dw2Uod]l${d-J1H] 4  M(z!)>?N AHG׹m*t3IM䕉Jp'w-8'X \")zz\<̘:q1Jz|y_CFqnrX(Ly'G!;Ӻ)z@.t\xWW}x#&"+9hFyM94'B)#7:زhH^1%i{K\Dsh6NU^r7lNdE(]-O/&%srY0^~^3?6:}8v>ɖ⇛ՊJ$d)! ,Jf䊣~_WٗbUO}qvi)ڢ̗wH5O -W"7J &"%ЈT:! K5YcCZG)ڧhw)G&JRP"Uc("jgbnsy膴u )@hAWgԬ[0q!).p͏KQ)m}dp sg 84f mTgČ{ukAY]H]u̳Sj3"#Ls8w#! ,yhv #ڙ(i!1V:rq͒嵟>:s ld"%͋2g~+l?fe$<)JqZ)1 ,IFITljؐi)i9/9PgABbWPv\~tQ[^EM Dg(oNt(BtF ŖKt(6s t.?Wh֫_A4w.l:7٢Aűќ0' uMQ*!` Q{mq G0LC2]ֲ};x7b4Ss<:eC@rִlOEbCZsĭcvn( t/[Uz,Rb3؋:kE0D·BP#pEm(FoFqk/%F"mh̿S@9 3!V]lGקqb4*g'6gQmlf'$Q&:|WUs# S]1r$X`quш$yɏl$!t :VV)c-JèGi#9tHnJ;#pI5P`>K'L$:sA%Yc(T1JJklgJԪa4h4wOCq2%]K8(i,G OU9`y%{f߂MX>]pOs V8o gsKe\NHQUٻϰ>+AGW eG Kv*'RG0]Wtѫ&ӹ( {1)d'))-BP>Gp'Jp4n)Qϥ-=8FY$L(HFfȩ "HDF"a RG?b}(5݆ C '~BsJp*XyQ+ Z*qj<qnC_?_?v,lk)Y~?[Rʭ]b$߾vSnLp*l0 c4߱EliMワk@-@gqp;w8L:wV ==>3146{/$xxMSٴtX5;+{Y2^ :hˏ|OW^Ac%x(jȫrz|~:#d3y-ĝKb60X-O[{=|w g7jf{[]ɷ7-F ^󾪕gm #!y`ʂͩ {Rp)[gx'n3V(kLW-Z{V3i-0ns"v_{I m"\hMSxc}B>|S>KDhTZ(.j7KG>3 yC}&W^[U)T+l<]=!xM;)~c~: 9O(xM\\Y@O?Yi'mEhwcp^ge,7h2Z.+>{(BM-):H2 i9KY?%ɿBB紹M7wMgfk=Ζf=|0"F2Mf^1Gפ%i^S@SD"tHo%W;9&KRD11ծ'"%T_ǘ Ƴ\dGj+FI4}^>wD q-NSחk`&$v 2H2[-dž1+L\T"crLS tr)Kk}:.> fO]M@rX=ĠE+Ss7Gf74Pg 'E[OTx@d徖5Hp/"̖{t4dk~va +nlu6 T4qF 8e=nA&RM\!B`Օ:Ą=qs4AZGk1 T#(b[-@v5NT)6ry@BugAW)~뽤nc*A6snܞɂ4FVk dp0աܰ+b%S(( v Y+<2ehXvh$} EDU/4^ɬ=u+eV_!2aP`"J&a\*][0W9v@cَ50At@Ots>k6?~?-g/ +[c\8bg2zجn $ܚ6}߾ ߬{t&@8xOCk>!7N*~]neU!DwVaN*\˜aWq^ Fi̽;K%&ܒ[M 8un|>~oUVɭy5 0/!.w^q|O/+b~Mn ?!0Q1UR`>Ww>_~ )BJryE܈WUR;捜dP889s1a0|8ւPjWLhY{mW[t2 _q]g@Όu=P20R ˽ wkLً Tdl;)].n*"#&=ՄB)#i%Cħ|Vy"t^1Dw ^QёfjPWW(~sOR{1 BWRth2o6T*d`>FYSY17D_ AoQP|6Qz=¾*n&sr@+yiL %1EwV́x'0p w|/V.SJt ˔ijTF3^:_BaW8{M'jf`A(JL7v蒯.`swqe͔ K%+WtDb :{1ņu32]MEZ?`Mj2vx%NK,~%۩e{PKWyz&ӏԃc/(J;xW֝,G47ʧ(()W3 *6JJ_ǿn?PYc9nl5ςhny =39ߔ\L{~t#WBj眼~KUEx#K BBa45TQc& nuhۨ!zk;RBU'qr(Uƶ ޺_8g"OkrNx}>΅8hezt I1xNsr2rbS{l3i ²K]㗇7WƙbD~_?&7CwO&~smq#+hO|cF;hLxS@h7z9zn /X03܅d0{7r#`r 0`F H$=4##Al[lOkU7AR2 Vd TΌ9an٣]ẀXD&lo彩B~rϿ9f߼Ow y)18oflvy__}YPLxݳOK=.x:plX3Ɇ~:PF-EQ51лina" g]i4;}iv`孿n>!V}*eeC|L`oJ/fZk5굳@GkgGJRI!*38yj5(E Q&b.T"]Y**sFW8 J|sNMɭIZқ2+vYXN7H8U"nAM-ea));<5h-D@8ȟ$n֓r6oJe\4:QQxBLi joڌӻ|Αi헲'qKoC9 S&4Nd$ ƙkz5xx_*{-">VW+4Ey 0^Gx5S6mQI/$x6n%Jh%ZرG8#/_}tˣyivSP* "N0ه7.7j*ͺR UK=ҟմ(LUi2}䶊7RQ/,$zx -]L]s\`E ֮Q}KL AYDjE7v^R[j7`&˖-dW 83K(fcɒ"R#MbTG8Xaa{6^?MCoz_1LR 1L5ns^ ERb=3a,=B1% #o@9ԂPj[X 9 [ 4jAeh#.ÚRX؞wQf,4Kُ^,K@5L?#79<3A/F/:Jb$6O1{Sr W94iz\_A%%0\䆢x'ҁ.*oq$c |:lYq9MWk7 >+LJ=Ewaˊ-#-827 BX`HnqcՃ.aVo 3 gꥉ|ϲQ )˜C}Aˉs4Mb j h#^~Xѿ[cWC\ޅt蜍/#m`vs?̋*\&X-L!igH7BZ̀ k=.Hjb>.Ln 1 (Է7t68_-Pxϥϵ:\b~u0%YƠAYPo: 4d0E0p =#O$Lk $J o~`?]qOv5=}圞Oe7/+V]܋{iس{6@bz{;sK)031 JK%Y拼Γ$hVW%Haф;\%n-k ;J]틘k.b2iEIC6eFhc~+ՒPvPhLvrq+e-1jjVS\[y =+qY6:9N/5Wte|Fgx .y}^fwӺ,F_}K(&,U+D<φe%aZg(oq>T_I-={x7\\{r Npe  7eƷqDexo hԽW){yo?a))P9M8sQ^K3+# $+`\I4/W<'gJhX-uԍ?NNݲsdN1?VMavৈ^0eJ<\|jcGPPiFQH,[Ldy&QNA<\ ^b'Y-k.m|ſ9/x#śMWK՗ mmVHJo{+6Eu< 7RdۯnKܡHcWm֍6DIW,EQV_|Tl+ЮZ^8˴ kp:5~leM׭p>\BW X6Zq \~Ιla*F=OU*&:4{Ʃ|ڨb\S'A=YlQ *OU-O'ҝ`]bv*v n=+~w>%p,5ӝJN0^QEzb>r BVCahbBDpyJP{.J!Ȍ>K]sb7E60%qp>n qb`?~AH`@.D%8rVaMLs?'s^xMPa5+?xN)ynͮ7TbOLDJ(k pVhV S;ա V8`GZ1p^F>Fcq3efSZ؂3J\!kEy ,&ƚ T&f*@nv;3-A: .; +)3s[ \34Yfs7+kC͉ q4}9XcˤP%g_==3Q˱TVa.#AQ k fvOnGiwV ,.EU=x4$r4 +'*̃0prtWpjG 06da)z:G5HΆ#XcO~ƳԌD V>TQ$i,5儠FY+\9)OxCi3c̬r&*奯 , TX"oqfKc_?Cj) jLĴ:x F sz~B2d;u!WקF횅oT2oQz2!J§@$B)pQi W*DL3-~,p$>ZaTqt6uD=,uC 48Է^J[m @i-08ӠRQCLxMbjY!uɧc}G AGc]Ym^\µ4s/*TER,Xו \)C|KiX?7Qyo@G 7I1Aamu.jX~ڿeD!^n[EAɞrמ0Zb/ p^)eĹrh8q^BD`(J) bQφM^ f8 g4ͺq!cFF_.'SN_?~?f\>邎6LymR{"fQ(5DQrg5֞"K&A.TY+kQ-iRi͇l0O xse4$ujQpQĔs0zV;.Q}Q 5e)dW!Exa ?Aρ]0j#yUW@\hN[`t\}[z0[QB$]F8X: THW i \}<;Fyc@\YUcyĝ(F]Fy[-(,R9$h[h+F´o6yc WJ;^{=vEĵNy_S[ %V{G 2M UԔa#R&t[DQrZ[!zvH(Νf*Rt3Z47'( RF{%4Tg7ŏj/*`Ma JSީib Y̱EJ;pCr~b N'@!fz顮bo4xS8.\AQX ++ l~ߦPKJi,GYSƉo}XdUќ;fnv'wO9R0nz:KIHT*wCU¤,rE1)P1X2 -fz;?O24RBq+іf(IffsĔ1Deh15D*|򊔒'+k&0'sǺ&t#ۯ_5J`"ֺp zRp$X=zO?O`4uˏoE/_JO~^݃X XFZ's4L2C:3Wa,ז2$fDyDITour'|2kL/i2#]Oz}L0F={YIJЄcɍ+ 4^VrtHӥȽζ3y sb|8ˊK_QAl+ĕ)-m"pAkc#ϸEk,c!ĜU^='BA1-uAea \z1^𔳤iMھ Ypñ:PF-!)}= ϊˡJ>? m#U٧ϰ 5Gc!_D *]YoI+_v> C] 0i흇AQMȒ%1}#TtpU%wuhV y00Au]-ڜZf`8Sõ*m,1V!zc 'fkq!\r|wZ>6AdON/eo=N]0\HU0~nCc]-6ܤrhdje{WZmУbkgWcr.AP 'H'DVvQ*ݭSɇc/$^W j@Ѽ]DxL rEԉ!2L& 4I OL'2ָXW?6+V 1L}uNE /Chn.D55R$L EQeRQsU@+!uV6^;V-6=(EI=W} a# &,1z4,bY)y l(~ttcCɈlv!/$Oq'P9ug_ZQ4W%&XRGݮ5v^ӿ_ٱJpN1˖qaլOFԦTaV,^sXZ_07h_gP;8OzB ?}X.˷jbDVlޫzh=;'ڧqsΏ]"w gQfmYTd6n5~ȖZoGiid'`-UsG٨&<W-ܻr:hp|؅CJǝ'LG " 7L3#½ub_ʢ{go{[uԯEE2i7oRB?8Bkn :zbE`kp&8&NMU4SPQRS>BR1G)ӊs0JZy4lz! *#yWbaүNj;eVB8ad= l NJ]23Hnݫ`D_jJME*D'TL:υʉɇiM1u0MrL82 hHs)[ƏV"=1Q-F$6sy~]AכEEr9:fzik Rm H/2 QT*ZH6 t-M\hs"3d0qjRFrxd@:!s׭gG;?i"LoSyYv!6|v8zSC=[.v3 W(^a;+Qַz}l }B,dੈ 9Pl5;+deTFٵ0kӍjE} ѥY_񬰶1Tm@c%&B`H(r9,ֻ7=VctYLgP:ԑiO5NzzYm><_|p|ת.CIt>CNyufʝU\xm̧gwntyP88N "`Fc W"I;)kGmpcbaLV>7`;^Ĉ{Z<}B3G_~u3C0($aAIEB7F=&@VpRvTi 3bW墁uB@;ʨ/.,54^A{ mgO0ᴍP<6ps6q?*3o/JZ'd$ ։v>PN3ܓ&ל &\;pc[|PVoQ(~FFyiJq~eW`0ّj~VZ`Cf=գV!5L,=ceq{ )qƐ4cꌅZ36CvR]khM3`ēNq*;d3<ۆ陝9245bOYK *ӣ=?ܭ>\-xM'ea0 0w|,`):7cV|^VWDgiygR0WPi1a˛DQZi}׍H |z2ΨB>bcC`(յ nw N)D3 Pt 2PIV z>%ϿӇWjd[~W彳tqPJ+)PR|_x{u}ðSS*)! ¤0̝Q_u- FuS˝[anwqeNYɕ7B dj-"0EmEgI1Gp2<XݴZkaƵ>Ϧ|2Υq!I(e2`zDJdj=ׁږ/1DO)BhY]!R>ϟ ++-wvaAb2$5om ֌H Uӝ=6,ePv5J[(߬M Cҏ&+'{;[X_JeaLp%jh<3B _v"_g81g o+IXBYjQԛP j)ԸW+:.ajd{`_Ik~@ΓQYnroxfES6ʐb>gBыi#0û|ؐ&iv;[8]Nڧ3袦}ŃEz کk !6X _N^V[OM%+cs Mx*LOȅaz0TrazG|qPu D;SZ S;P}y ˆs1^ߌ0|o94uS{R*K>".Id~PÙʩVOY|?=~.~5]6-wעkVJWIPٙGNHNS}2P6g {f(o}}TsA;Jbu_bw|bU.p" .EJkt5 P,tۯwމjh }zk/.}54^m[WZYY?RdCz|^<[$bxfƮg/q2nQ6OyCpW# }H%K4Q4`-Űrݗ7ޣGkz.nquAV~Ƴ&0·2ۯHףKBqgiP]9rִ٬`< ^t!{^}KruGv⏔ذ*'B~HG/\ehȐk#LEFPٲSyA /k(3puZXkv׆w>79׽ؿZuU=yT03sok;Io0S 9P}$5RrB̾5%3Jw F:p?)B64j{X'RS)Rɛ^AmGޢ TDL&~9K6/F{F2~wU/^r5do6d<KWxSMo3돋\DaO}U?ZLBtǺQL[c sqRꮣM9nKbE:G~3!H 1PX*Ɠ'>YoJHړh6uףgƋe(Ͳ|=);v!\bBňJnsX\o {h7ctɷQXdǧ-[_{{A}DjQqqm+8JEI~@z\|0y0C+ZaҝN߾BhPs\Or‡Bi Z,Bʊˏbŵ.ۆW#ҨԨ\%٪r&MlnQ̌5]%I"#E4rUxz|*@?>]WbƄ:$Վ=R}fgɺrGE_ r"&`Ga϶=K0j4p Ovռ0. "lUƍbEqD3RKeK}nbRm<6(uP $45PB#5_x޾yrH n[ 7[@o \,O7}[b_.0gDKNoQh$;mر HMOOU)U]2{[J+?lWR:GבFds"8;D6ac6x<95 {~$%c yQl!{ D*GY* }{gFWXOONLnGD3\yКW29'%r -]:ZzE{Fu#]@?Jm<^?'~rw&ʗRK_%6jHR$#N}[n9CruѤAmɼRmRA}JU B)J!iJmF-5c)mUX_x,C'7156 \i`oJHFPʢECXFJ=qZ8tFs"!D9iBJ08ô_$//|_*nVvʵۯrL L{>+nP&c5EJ[vpv 0 Y 򃏼Eޢ7qvn{E@;yvoOPb=r:pם07֝_>d˪9֥:DR;ɐжeI3X\HbU0豛ڝXaks2(1.ɴ Q1c4 *p9&Z19[L7VFsKdBOPPay46yNG争1(Yd:cge(#t"%sFI`nsbWbm7s0m.aXCh&9ʶ8* <J[71 I6\sff%V ¢*dͭw67b&$Ұ[՞UʘaYcbL{mt 11(n`v#TS5 ~|ijBU6;fu`fLN%r #May@fLx0>U24ͦ&=K~>91oȼ~a9''G3kFFqon*&4C0a foM=&vWt|U^cpz6&,5a/;|)V_ . ._]>9?>~ QyTiϧOY>Პp^~eR>^]Ԡ+v/LLhd Լwd^=[y"U?WKX }W6 G$μJkT>vaa$0ł}^6B3BYEYK/9΀h*=ia)hڈ}1&i6uu*DvK# 4YB~oяy-ʂ*_K^s=.ܓ^5bC y -?U œYu*;DTH.tAeyTC,Khі{Y229/=̻(4̈^]<~v⪵}6֜Ʒ7J|HTUǕ[uѦw [*SGYi~?\{[Ң.扵};7;=rTZnfTQ5f95`Y%R{Ͼ0xuk )PxoѪko~;H;q8/N "6QѴP)# o>llj}g"QQ^GlX}rS辸wcͳ. mܔn0%cYM֓{waޅvQ`ydo47*Wqԩr>0._7i> 67}SMp*B>mYSg"զO7}z;=}3JUkQ#$,T3p#k!ΑFVycD(w8@8F"~ko~=P {/6 "~r#L 积~~ܥYᦠ-䏼Epa4>f8GΞѻ_5ڌW^Qo"ORB7y=N1F1)ts6GN~#7)KiFQ}8Rzri黂KH^6)@[^(NgF!jNY㵫Rb+qI`lʰp+3P&Ň\SUƺCyWBVeH@YԠYCZb),*tނv\jc-7ɇŁ7Z([{8-_:HӷR(yfo(ATuf;̵J-˹ ? it!Xg=s J71dS]<W/XǸxѪYkCDrl(Y|N=H"F0tC; I a`AIok74P(җѺ7?G *| JR+n~[b@[+o%"Fޙ+Bn{gF+m-"uExcKK5CEX!$֐,mDSHAwz 1*&B>Hv|Nk*yksΞ{Lm ȡC5--  상Tx&Z{UhohA?:']Iccbkk}_}lԪsgx<G wjP67y8f A<Ój3[sfGԯG/鋽`LoO߬{ <%ml N'ucƼä}W,*:p3k} %Y14b^%DF5URA1\\yuΕ7- VHfy{#:iחuo}}|rgݺoӑY t}g(9\uF*P=Զk>Wj7]e;Xf=!MB*VLGkq4Q>G(qX{Ucm RH8^v`J@Ha-7{q<ƒJE{bķZ#O 6,䑟=ipv_TB I4GJD6͙m nI}Y=Ng{;e~c­%fbL%ceaJjdYL=P A= s?`"$#HJ-V̧ȫ0V#)fh?WDT\az=+ya,qpO@D"(0sJyDͱ{.l CmmM Z檍g\Cl6|^ċS 餥ڒ#U{gާ/Oe2j$Js{FޛD?Zk0@T:FŴ T|>ğH~L7kM&}iw7v_{fwWTnx)A?^͌w@"M/rSw8VlfflràXZMLٸCOTAȒ}E7B9hΩFsN5sӂY895=mծ]j|II&@N< 4[Tit2{ju)x@BOSaJ:hh́LҳŽ> hC6&9WB"Jî#_ MjSEI`üRΞa;r'۟שּׁ_]A~U`2y( )e&mMY_%Z^ЫI=yXa3q;a. # G1oL93KWo~DSDH;[עœE,_HbbR}#T 2Hg!}fI\3w7GFikqr`zWyMy(j(-M8l fc}kT<v`-U48Qj4AHADt2%Z9ͮЏãNTmDڃ r#w4}8ߝg1Ί?nsr#-(aVP)8tw?`wgf`\ ػvV-fJh_8Zn|hddc.,fC C 01cArUXjV,Z/ <)p1ɂXXs(ێ-b.F95!QAC"f fb}]sV7/(~Ż79k* XLSbIjfhR))$j uԺ-F-`eFx4-cnduhwڬ~Zq~:OmIA>boI/.ێo}02ez9Jh?7w#noֆ\EulJoa}RBC|T+LJs(wF3<]\u fw~}xn=26E \ ZV~=o}O(/+ zk>Fϑy_5 558XpJJK7M}mflizoZrbK[ju! >'YOƝ & EJ2s-wT7 ;&@vI*JX=D1hsKR%B5UѤW-B(s94מH)G0c1X}a4?2Ӳd,?>9~}  :iM>M39yt.60Hv̠K*zm%_2gO@tl=t:o}fgŪmRރd]5eN,7)Okm$w@ 2 ֙Ȅ5ku$%ra~s!{ߗ8̽ye8?#Dƪ>.cIYC1$M5@-\L|l]R/"4Cfv>.<\{ES[GNR21:L2B*>wAMځݒ$b4oW5@S,bΝw<<^fK!OEA>>RκQ{I^l+7(2eCڂCg] 9d(j#Ͻ@:wke8FxS >Rou F5Z0тQjmęM Ș@( |=n7#]:,Ny`5GG[^rM@ն.TJL59Hf؞1N@^[|zdԳsYeKSgfxnhy|F-R+#ӋUybr)Q`&E [(-N{3c)#kQ91ݔCnd)#$G%17;c"BsC&Igt OA=h |!%΂䥆TBUX83Zb,%˾Y"?_y'(6/ .'Kۛ%р{gy2@ѷofΖ{2O@'](×B7 epRY^,:gi ee{g#_;ekdex6OY|P2<`"3 .5^p3ɐN0?C­w28O4n2g74ڐhC*mgհ;Ag_ }r_ &F#304QBARCna;{Y7; =c>06 =dSEp,v+[7^bql&0~1ڌߍeA/0)>]KXTzڰnxACk6Σ~ڐqPz' Fi.)ct4kݺKUK痯xt]YYKm xW#w6Mط/~^hvJ#mdzWף ηo6<15֤ qWNn*] =f?~Ĝ(AJ_zׯt0҃587؂x7m'lxuO5s=]p.d"^^oK@j%D綫K-YKzIxN of<vq`h3!XY=`.L2[,8I{JyHrwp7CXvfX(#V;ɖI7D/}Ѵ/郒IW]dz<]o<:plx|zwv4 p2?(Y*y*u}̀cÛ_]TV,Xkᣋ3%geGjol̂)ZvȓY7d-F(ek&@|Gʻ1)e!'Fw ;\ercEj|W:RKI9w{ƌ6SXߒ d|!|7%p(eMRɚ!CJrk5Rj%aMHM1 9j( JTX=^ 68u{g?PEԟzXz؍$сBcE"!j-~`&O1sbo к+ }P<٢zc3 J\SeQ-9Y]Af.c's.&0k*8;烳8lt8Lgރî 9ad7py@2 PpbIPAp~?yc) $"АL1[xw4E=ؠVb*Q@ !0 %(W9Q؞e9{EDpO"=g\Ytҫ ?muWbJ(p/nWqȍÀ1!es1LT},rn$؊RM*ƚ^LIs{I>`m+n߬/j$=FyX|Zl:ghzL$F^Q4BTY$pV`b g'NTx*֧"5c@һ1bo{pd|BLA`땴;Z88nϨ!Fђs&r%%1Q0r-Dm#ejK꒤z@u! ?>\qX>;Tjq)C_0ZZnY +C5VaEƀ`/-&$[+P1Wi6&v4o"U&kYȧHԞD 9$HV1QrUiKIp .#r&s0*b£wxh^\(zBl @[kBAP-$~>>\nY/97L>HOEGw+!}ҖIZ5+ӵ#~([Ar}h*v+az'9(V*]PѓT84[ ( jn5"3ԃn'eI*Oyy!"O>r>6jL< _~-hMٽ4䘔'X8 \[&*LkJSM-pH,AҢ,Fq>no= xxbE,UJ2(1RRd8[{A %熠|HNtYf1B"]%Snc(krNBl0'Ӱ"c8*0h7b׋V^јW t gfOO2"ߓdv{뜦fܡO> LX2hȤ"&*CNze qRT#j@)PZJ$4"8Z s.)aU•T`YA~Dۊt>!%8^͹Xro;UF!;X%)/9K|#f A%Ԯ%q7`e>[vE#GPm֍4o/wg>jSYM½ݶ{o}t;;uYհ,Iբ'{g?5Or1:l}OәN=Ry# noJ?$.8nrgOҋ@֛0xkV!Ví//v`[?uj9zfٓŽU؂[c(T^0kBq2 #ڜq<'$ƀg0`̾ڟ^ˡ'%sSE$[bnKrCȱ=ܬ' kI_۱A{iM!:W[XobIl)%zPx%xA ָx1;sQ+ܫ!Hۻta=o82$0 ~|m"ksS"ـ!Q 6i}ki]'8"3=YbItw4Nz1bKk̹jЀ_e Lc$ ~I z'o׷T_. //ap.0scs-zl,[ ʰu #FH:^k_;'Y#`34@6r[0ӿ78~{)*k_ )x `5ǎdd>p*gɠ7DOQJ&D7=fYxs!yN$Rk9i&Q@8q#&Oy3DP5PM }k,ݑC1gT1LK z5XGkYf=72b|E1ܧ5?oη(VPVwY Ch (,)!ܡ=5 ZIrjW=>&T~oFGjڻi>qz> J'G]c{Boצ?z&#=fݾzuiPkr(^GhD9,\mW}1#@pd+:n8gvFzxTXP@ltt`Z5?];s73pӣgMr_/޵>x*uKD ݈eыG p^!ȞA5 3qq(_u597WҕqqoBa~r $iLR%,vb7}l Pv(Ue0iGӂ,{h7&߂xv?9cT8󒠝?d1g`N2@ rKsvzD ڡvAq]١X뼼$hߕy*Ղս4h9c[勸[I/|a=F#u#u4/kmnFEO;UJ\ݤN>L*Xة )"/%).?@{-@ JMn=z1]ؼG#TQᷚjG@#݂=ZH#;ʋYtr~V< ck5h!6i#Ү.P j9ͨ1#xVxvM-7Csv(b)M?⯨욲{^\['_+?ݸ9U,*Q]dZ71u&݈Pe9d.) REd9e FT.qwVFy_4gU ˢ0jij~Uպt>2[8x Z.SJ@\.q@EYe[GV,G\VS6XUI1;i ,Ր;:H W,btMwo֛jJwۡg10qTk[D-5s$]B7]aߕs&j'Sr &+Ř`-'6gPZ$`;IQwENfF,W\%[Fy%J&g*) 6Х CQJ"Pd e# 4YRT6s7 %V+2lJPQ((|= 8>RURww'$MJ!x sB1,"9b.jK Iߺf̌REM^a-hKqdk7~?]=oK ͧ,(#3F.52T <-!U B&[Z ,XdЛteaV[իQU!~U[i+2Ivl2MМbTphiƒjM5i̥QiONq 8O +,&횴{,VuLEc9g7GMo94zQ15HhD!R]o–OEљ+jRjjiA&typ!ƜSt۵Q6m66dHXҐp$DI+epSORH"w4ݝdJ'<70ߌyp9폓+&Jo#KWUfGKSW}N|ǎoW}wnSd)؃}~6Ck}s cv1jl:]WHXU%uF:}>j 2X%{)kںNkdFT:O/O:G>c6c#X̓N)T:)%EhϘ̺_tuqJ<+wܷN't- S}*:ghDO݂g[UF<-jgi_vlgqk;=F;7i5,ٌP[YSZY;{4?OmăP:H>#@X|6c$Pl %5QwZi.w$uA%OsƗø IV٭{2Y'}Z hmh>-p*_uJY˕`'|Y˕ ZMu{u[ukVoVhl+>])(TLJͯߨΧh)mB=mAkG"Բ-hw 4QOȼ2h!mAߜ-/hI+}Z2b+<"n;ۂԧptIݐv AAv34g{v* U,o®k~v\!jV2$jTE+A&>BmTKs+ŽЀMMӜ< Cs2I&zMS?}N?NADy3jV}?m3,&vDڙQV&ӜdT=G-imѮ)^z#z)I1"q\Ӕ]SvςS6ǁ|~ד a{ߤ:/{&%>\&? 99.+)_׫ VG_u|z']Z[+b֒Dʽ0bT-t[laSr*(mKɻ ? *j]:L퀄 2E:jO"$ŜrGI20MHKR<˜d-CحvVK;+]X-ZZټz}rI021/Q ϖ,ռo{*HO)*Lznky( 5h1(VL/x}wR~}/OF,HS͟׳;LW xYm?aFg>WNsUiڽx_\~8ɿGoFegEv><>{T7byץw78A 3?7o_n # k\VH\:ZقH<\o?竓7Wo&=DxӋQF1^gLxdJ4.dHt`(F+0nDpjbely֯&ROhg]S/ex)mA-̛Lhu!9Ô8ug9SfiX9 Tm<@e?'_WnZƌoWKK:n'=Dx@! b^\tb1KcK,R9WTHvD?ZBKgJB ȊgeUP),K&0`2Z4Z#PSB{&}<) #3Z'0"dVQHC$ 6kt&xN%|x lrHtILΓNA7oK)'K%6z6yVi}5{EVIbI3]!}t켏&xYxRUTW@) a!j`ȚX (q> PJBCAEH^)$錆bF/EoC m65gǖ 5O}],F@JP2ƒȲ%Xh)Ң*ZJ&%6UFmI' %bH{lkA:aaА 1 Vb֐?̭D8BvEpry3:#zn&o1w{G?OܔWNO:nkZbzuܘq8{{$Uk[6Z<:^<]^TmxgAu«8Gb Z"@pgF-_R _d }OaAo,OY[KmqF={gm۴%mNpSLء7?:`Ѹ |E\8ځPϾ ovt0_?I?|[Al#6_ݜ]Ҳ(2va#^vl{^.ݐ&X3ee3Tݤ W@f- eֽ 3 Nʠ:tt.eE.RgjGQvRcOs?х\-.?.exRXqo9sWUn[7oO5fzCy>xnKL}hH#`F-Bȷ4c &s&m<XkeU0 Y7ub: 䉱ȦX'KgӋOG/7-a_'3Ԗ=j F>󹨅ĝ^_'wg gt'lyE:9.h$9\tcIL~HWf, K'17b6XiEV 9.+q@Z*!wzTĜ-bv:B/#†ܼL@3 p!%}\M,,]U]U]Bb\J!F+dp*D U w*ڀ*PI#3,D-{@7GT!8%r¥Ej})#94s.hW3j=_8ऑ7sΉž9do xP46]ƩX;?ΪLKt{|Jl+ԩ\ r9`ӏB>n!v|ݦq7VBx :0#e:KsR Noi5j }^] CDt7~aANG-Q=f]a:[_{J(S}V-ΚN+`EZvo>qWB@XSL|~/b5P5]t1̅qkmLT.JI+/)ɹujۜwe.%#E'wL(R|ɻ?oϗYZ2w.MVTl/<̙sq;4;+; ńw|Rz0(N?i[f^"e#*[[U2-kLZ%* 숐deZ̗S,ҸݱXKx/$pﴨɍ4W&%="-, &c,^iI3{0_Bso&E7y(Cmz(=G ,H1ctrLed4'*Jn$ uνҁ{O  H3[h~$q 2 8@-/Uht7T9̈p'"pd#J7я~o4GsАdT'ڤv= ^rǯazQ@L pTXă(gSƔ,E#u2a{_8(Rq؎l '!gGyz6݆ꊓކ X_pЄRv>wJ&W|HAK|C)c(:-iMNnrZvN˦y%Qqc"[B [o"9 Hj)Kup45Td(-|mj9P{v˙bL}j@+^IrΦv)X 'Xn_<Qk?hpYgcm]WʙK7L_sĂ]mzoa-`82P*Tt%6? yylb:o2fG'ݢ 7BQRQ_6fo^УdG [&'-?ZabsQg ~+͛X&.%BH"~p͂:,•Es뚰 }ȇrzl3oLONļ[Ss%m2#T!ʝXI3.Rn%2dޭ] ͯ0QV<}jt1S@p>Rn4OJ4\l19 R`j(6DFy'bhQ,!Ԩ2hČH R0S]ⰷc<P$22"s;2;( vȨ%d G GRP SL6j o澪s>t Ko~}&nd_E-0:v(O0I4ӕVTHA= eW.HHP5> 5K3AjTlaAӒ1d1A~$ŔJeC± oet\r}^#r^Er㛟dT /Yγ͚Y͕Q]PC Dq _U$. >d ~@OOKZFUx;o0#9< ;s`/o?< Nct^LP=FJfq/;o%-?=Yf5S=A٢=V*S捞[\kvζ߹YE^]Ѳg8-EwOO[Aeqژg Fd2}Sg8G}y#~>AYU6{EۛX \2F'4'24 # j9z`y b>c?W3u#!n`M<1YFbW{qZiYVu+Gt2r8N޺*q_KG[`*3N cо[_+u43z_ CbmrԴ=`HYtЍ׃m(^-U!pwFqzo?kR?|3y[rybaA bס{˄ o~%c7ph?5F!vxy@KZ& W K!ֻzakD SKv[uńU ofYy+Z X YHO Aǭe&| ]0ʚ7PrXkG=Yk\9xV` 7dbkR"꽺@lxzS5jNI27AՕB ={:cr$0v}R5$:i9JQ"bN>6BNASʂDT=PZ@S~҆QPFlOI 5@VX^" M+\-f rõf`P:J"MPV:jYE]ғ*Zc9٣(ӟG@j!:-Kf*~&D"$ain^@smϻ 53f$l-L+w1uk"ƊD$9>*eeԩ$LZO(a$;[oRBЭY ̯E dA\I@6Ve ^έB*)JSEW)9")XBյ .aXGĐ g3 5ktDN*4v}4<~v4&N&=9y/\]4|ge/v_a-$&FlhÃWRF[#*hd)GRFNf%$P:Ϝ1kZiZu6km10k7RdH2Yk:p#\;N0Cf4-&5뱱/p/jɎQ$-s9f^6J%C'P+,T "נaS`8ET&ў)r{i?{,}Q Zڪn%~,zlq `Rɍ}< Je9O>kgII[ҙ(DŽἅodp*n(i9nlD;/3 9#!#dB˪#CD] yX>VtsNaqJę;Äh)PF {t/5d4ߴ2R[e/Gb$ΪE&6ST#G*¤fdK^)c)QlG6m~b0Ώ4$)eL}K+ԩݳ4hZ߃ O?VkL&gZ79Ӻəm:&AﶲJd4" b3 vF(8G:RpI4w5p׫ΫvҁPr=z h@?1;1hŃMr!A[֎+$auU#QJ2K᭷IC&漸L,dQ[OKrqb`JBQ(dXx@u9Mr ++C3%&wRQyF7c8keۉ q\E)vTQeUDQVtz#Xptpv VIi~HB2eN}*S&J7TMJӔ9)&.ΉJ"D[bu$ )e2K*2e6)kL>DV|>1p5aeL#1aQ$H8GJR PoCEG4Q ̯ Ll: bKo9F9j@Xt1Τ >50><ݹ~/3"+n9wJ)PQ*o3Y-.om_' `-l9d&[Niij82_@`L1 b l"BX/2"i02%caT6 !j yҫ1׳~urRXwjk)y+(꺺Q,B`}hwvoޤy1 ۜ@MOVo?Gͷy9VN?3;LUԨ.]`xC 'f:@DزhU>&2܁V\qv)ʚ -+"c] +cհ;deL6[V畱CaJZWG>GkAzAh|qwa&xj;#Q-,95կ N쨤s^5F$,0n"@+)LW0(E\2i# X !P!0|{b֎ ?֧|c(LyEKJ #"mʥLO6^8) R 㠩 gR9li;;5BOgW} ![jDZ9,9_WRBWgAZW 5hj(ujNKc,1 .ٌ!fnfFFfd->`.qv0(bVVz(m1b 0˹ u N"u_P+ lFT36#n{n3&TٌxfDwZC:qƃvM??xiktڸsDuA}ίR$VG]܋.s6[/gv7dAg;vBviVJ#y`gP8M'8!;[ܺ廘9&TGMFK}oĸb17J Wg5wvJjnK%f 7hH^>0yhTĎvL3z-[L"`~0 9FGRUr(Gl~|Y̻>Oχw R?O{/`~{$^#\)|$Ki4-~{O5?W=䲇m_:0a|zv"2C/P{F%ё{1|3잸Ť4T|;OufiF۫XY_oY\ރ4cG^ͧ7kh0 xvigO߿.%Ϯ.}ե a}z({ӿ6yx9_}DLo >/@؟<qS31i(ʟ>q0]<~{;i#k_=n>~Q݌~{?XYܗR|Rg;O{mr-Y9i*Jp1! 9秳LR Q9jdO&Pam|$}hC ll˷4y>:8rBSED*qV5\ùZTv?3kᤴ>gsl5*gslphe9\%C0_N<5R,Jy6 OuL?1Fٺ*d;g+Q>S^3 x k]0:4j0ڽ2N䱏~ C߫;ǽ+--{Wyqիo;Pr?XT=_}r*FKGky }GшrAV$&0 g9&qb'9&qIتe'*Q J,QZuPqJ%P h,Lk93\/WKMJZ[s a*!l^a``ɟf}qN$Am RT*sHjR4Ysk&F:xRʔ\" R1\Д@AHM,!)0DЄGVbO*хœ@~QvAl #1 N(̎#:PZʭz lbT<5Ob4^zRUdߤSKLYp` 2LJ0PLHaKI +Tf:k%|"8=p%D BҤ"X!Zb0JT \H:4$LciRù9Aq#NtSƅ!qلQlcKn-p7LaA 4 6=H@,yq*@a]*nq<( h9`f :UNUzN N)0IH%KɁI7,r\S"TS|)=s7:xic̘D8iHU`͂e ;"kS.`;+!5)`Lx #[./}2!Do$CL셫öJ$9SۃTxçžh7*e ^<-e='Ȭt|5h&*%or UQհx=4GҢۨٲ9W C)[bJ!u6jlY\v4@;\T)xEDpĘx[gn0d4Qن0j'k[Ѓ -3 ̃t`"oL[~f`f Tq۫<}V2< ߱]4Q>|1$%$Ԩ 2:l$c{g4˹yLHy9ל+Jd0$dFDž=U( -2@rځ^Ǖ\S-*1ZkPD&i1"}ls@}`Sbc#&O6β0Jf6i ) SO/%p *hF:iH64&m!e 8D Τ>NJ*F`9G`PIbH&0"˥`To N<8ĈU-CN`%4ڕ㮒 rѰKYӘQcN 6BU*o#Gr0Èǿj7ճHO_;f6_bS|jP( jF-&Mb`2"ѩ1 8MWKK$6cAΆosYZ$>_4i+dcഽA| =m%բ*Ce`%]i<{>NMnB#>vLL4,Ove~d}:kv:ok!~厩33qmNd*F OپNmmv `t<|A8w ʛ䒿i?dOy6|:O(a<'3pidZ" MH x; '%!0uHZ^dɯknygN /: UJ8L8uDR#U5=O$PIp IIT=ςJ+Ke@ۃUI@%~Wۺr\_7.(']Qj`cyo*P',gDw6胑ӽL>2tLW"!(A~|ҹ Ass*+NeV!C7D#v\ %v>A^,DF+_=?ų5w sإ'} 4 1 Cd S^atxm/2F|^]QgkF+KFpx):J}ڳAq'EO^>5EY_ ˚cJ4/Œ1%5E9f9ΰHlY ? Eľ}`{wҏɮY2nšq}[q6Y M(%^UG-]Ͻ)lE,b@2)Akq*43ZYkGT~2 ؊⊁jL:Kb skIDH|+_ AOȻ:Y"Is~.I5aWs+S:>\@1oK W3Fv?e]<]7gri<^|'0ibidcwRL ,.X/^l 'l||Mߌ4/% n}?kmG0g^>i~u=,.WW!R|ӵnȧwaTdʂ ȯwA>9\S~/adn|Y+4KM2hmyi ߖ=|t KW|ZS&ۧMH=+gZ(粣& *9 Wܓ 2=!'1nL Z+ZVպ&Gu]NFZ$ ƥ,QFRxЃxHPr7\î:yE լVڊI~$A.cc4t~^H^ i֒.pOHzqCb>7^WW_yiԫ//($?|%T*!!Qﮃ{X+5p#G1g+N` bج|G~*!<$v9==/$-E#4yHjf ImΎb7ceF<_dH*BR=hN5TM\ 2cG`AItK2("'o Y%$;XōyMpI¥5 QUsJ#&C*N7 aPNԧDo۲+N&pXP*'w[+v Dd}Ly hD^I-(֡`SD/+ JUֺyIVH+uϼNر@c >x]6㤓LW\\,){D.@FMA[s _l 2n@w)UjJopA'a}$Btҙ]iDk"GY΋买?ẼL~`sږDY]ԫ;sgm)eɊBnvXɸ*lm˂!$R V:pml}DqÑgتθt&Rp2B _SRN@oDkQ=?φa~,8?{P쮯x*$vO2S/YzyСҕA|cIsqPXTzUT%u*$z Nj=Gqsfb:[iïGUwc,ׂ%eоQ7澏H|LVD,Gw)Je2g7eΗ>}!1_5c ʄڕ{TRp?~DMe&@-߮Z&W +zGk.wOl-"jˠU[^ROY.k%(/~(;/RϬf5ͩ\(g؉;*{Y}~nMU3ƀp27?ap= F)'7PʤSJFȯQߗU xVejdt$&d &dՍ>'w<9S*r=~cjtvdToDLwGt{oݧf>4gbHwhܤHwiûw|ɌdƳVO?8=mwLcڻ`6B Om>2gnCKvhz6SQ9GPf) R1juʓ}lCÀ(gtqo[{҇;^9c@fxXĒm=J)rK7D1H?ol8&M,@M\픔_tYg24;/sFRp=Zb#1 }7k븋Y\l>7 A5 UQdj!II\'XERjZҚVS*o,sӪ2\&D 9]D[`f[h DNn1jf/``DY޷IE` ,opzu*DIYt[|QS|}f'W%(8lN=uV6z7u+LNI6]݅]bwO _)\EygdI_O ɲi4s԰ ̘.eaƦDHw8OK MiwFHE,I̠.e7|~5GXjmƞ ] P ځ Tz2b`7~AiJCd1%DUQ"Ĝ+X9+`ii86$c&w;[Mm])| 7Mc0c,7-d.Z8 c!$xgX]\uQMW_dfeuFώlkG8odi䖂00܀+YP`fg(ǟۛ≖I#rHPPuUxs`8?|d(Ht,$p AjrHo[ccr} ()@YD, 9L + e2#D)qp6>鿄H(4N`QXmÁMT(g@Y/zF0m^J5&1GfrXm. \OW(,n.P\|n@W+zox^\ܕ>8˫R1K Y‡59?U*D08/Y| o&D m;O&eG-nr$l?4D-A s/UD騫;Y͢TRUVӛ7(xl׈?ס!yo?EŤV2o|қCWI0a(Љ !ysx 1=)Ey0\JOrj4MAGv VPg9"z:z^*v9ϓ}؆yڍd9D7>zɌ97:X||K %.._;p3Fe\)FF ("- 'j]r:yHi+jz9Y0&`g͓/gsoVXՃ-kX%~7z]_%: kǗ(jB(8LZ(Q:i -"P&\ Za(AbCL%/׋i{j neiIhi[Z)w7!.U\yP4y%o&o=w+^KOLn/'~FRtmL}X\2k5O?OP`cc_ Pͳ K*7=<3*+^$+o.ØSW- !ul\(h0D3`x#3qo{5i魈!9 HDss2ίd颐Eb0qv15K[ksu d9CQ04?Fs3P;;RݡC0Y_fogGeͿ԰fTۂ'7E3ƨYXX6c1k co'k<._go:bpM\Mp+`>@vB3aײ_O6·>|suwq9[}⳿NV=`e ZE mݭ{Gq4B+"Jhdoc3ދfQܜ@+#ɡ :@>l\]&kz`W!ƀُU L3dP3W3dLgIHÁzoH;dD4ݾPPH53aaTeN mxJj&G/$ GTRB9HJ>DH1 XV]g-򴚵ipg) !hΠd"1L]O_LWNW#eVAL NQ#1ӈbD4FL#FI 8G8)lƂk=Ĭ_oZ,  N`6,Qz2=;ळFƭiVӗ^=)In2TY,,c=Jrvd8xӥ6ig=SsgS,R:$,Y$51u9B =1j V1, n6Aq2O-]%H1&.*gkX,'ŔK.ơBtiCoLkyPrU;F1zoL.$"T~n-5 Šž,6Jk(aFjKƑ +#h&-BS۝IS?ZjF*B_W$|؍P$`Ka7bJhѻTJt[-pƨ@S,0¾h`kD)<"v($Js%z[Z2~[3Yd!k4{C?mzk*u7wJ# GE'$}&Io)"ABҌMt^>z~7 M!K,oPlj*P=r< ; ᝮ<CZ$WB3 ; Y I$ NWƣY%'Z NO4P^!7NHȲPv%ѣ5Bٙ@55+b>]#^Rmx@Izj7j6c=!YRR!J$2nG1FbA;wO3Jn&G d'OnN]pLN`|@6n~}_bp᩷99-4h9aUvjn^3$41*0Zkv(}<,i d̝{@ vDž.jU17GkiX/wâr)W@N<!r2`rw`o) aKm2X`ƪSKl9nnc_\3Z:3`Fg ґz`3y%5Kk8o[SCQgASqC#a#ZIւ $]!L$z^P¦bT+IF;eaG5+cA!RXVC`E;i 9z&a$Y) y)j C@r< ; |4dNKPrtCdW$szKщ2CR"С*mg Dc/Ѵʪp<()R22wu^}Eڱ*|k;UY+bݡ*k 󏟅Ȓ4xg\ E W2=s+$X U@Axl*L% 0=F'葒Zf 1k%=D.`Z:2<j$D̝\t CG#.1̛kCgr|gzǥ\AT~}zZ@Z%G=r0{-M%:5Z]E[E/ZZТ}r6OaSk.*|9=G(FL0"Tl/r[TwO)D52{xq"D;))ҮYAj6@}dLǠ-ou˔-CHbDW{=vOcD5A="zte1ֽ]ms7+SpkI /^l9Dqq*gT+35"%sx9.EӍƃn_`LfqGQn3IB n"RpN*Y;H1ͤ4uD1%Ēƴj٣g<No!tS=ZvCy/&/2)Nl2N7&au`}0D3$D ;%٫8?]2B+_Fcp/_ e0?b(Zꇗ1@ߟx ѹwɣst ]'? E>]m~m[g+0J 9r)Fݻ0nT1:xElNڟwwۆFԦ 8Xnt[,UL':Bi)u[U6,M4ʦ8̻E*h|28߱r.7`ro|5k &QJ!K3:4 LC&Rù9"Z-`2v,&F6Dok П<絼)%-ke3C)ifS8<3D0[<:1(AzdO0(C[}mզ!m.ƀkC]ܥ%Fm7nXخz0hƜQ_GQuX읣|נ GQ2 08ؿa{ h]v4") k}i}kk֠Z6K@Uֲ;gm4%Nq"8%/NAb=iw w @*f) o5NYsL;GCx9wD O+SxA`im䝟y|+]@ -+ՌQPqfrb>;e-5m{ow--VnŶPK"R,M=CfTwNH@/$'ZЪ!9۰#7(XnR2:Ju >XKi)qécN 9r1qz7Ay[,UL':p9Ab"ݶa!GnQ6ir4ZYM3eX3i ͜I=TRa4ds䤡8k$OS,50afB0)n"Ds8N2ˌ[s`DRc9\ ZTH,-+SDf1Q19"k4FG+F3L6B?AQǿX 8:ShA۰#7(Ҩ#ɪwcHAӉw; ;r7-hU 9r)8W*x:wŠQǻrAT{nтVn۰#7(*ݰ2"8ͤB>LO8qA2`qф8*uFcY2; 5yr^G7t>f7iTf'O` h3> az혞>9x鴒+3| I:(z+ϹP v];~n%ɗN œtE|)<^9kv(1/ u\ƟG<1tW?B. AnlyNPppgJuK ȿTz0 ƟBv^K?I:/cM?n'h<2"M{=sYAC_5BJ )՜#y!xüe/-{iK^ZҲTg/ѭuDԪ aӟk32dzN>K9ծ޴?-iOZoP0,!H :yExyWٹw&O |’.;y|"Cs%r} *GrVryV^>R(_Z\&~r? L+c̛UPl0Ü'̟0v0 '"%E ?0K,צDr y:J0<}t RxYN }+4qeΖ(㢢;⸨`NHd^&g`o=>,=F܍`]v:1 xN t Ք *pDJޏL:>@?N6Í?^<zx9|FlNc:_h&q4CRE.-@GZ, ?*\AsB!J"h0絹ow}?{W_ďa.f '齃fsGo92 s|E>) =NSɷmx2v]Y˩C?'d_^o#Y>īd:LO=Dz[Bq,Kto$7$6'%h!!f2ͅ /vOsQ̼˟7<%p' D xvsq9lD i(3VHQd5g)O9I*K]V&{by L#VsA: .*T F Z5qr&cSaB8H7"K,UL':m:]D2TE ZջmBDclxƹT3kbPtb9J/hU6,MtɦpE'XbYXgxWPfC ͏?<+ (C̯ 2)wSHSrpe DgIy%8u"-_[&3 vX20 ( @z]'OϑqcLiNi*')I Ji!7C:fDQkx, 723c<5f{Co,gyMґ0zC&X♱pEd0V2$ 0H"-_(wgRy$m9Μ[#佗0Sm1^ĈOцwB`Add$J))hGj<, J4,uJg1%Fy,ZV\c!br&ܚQ_$9K`bO# Zҝߜ8a]f[NqmM͜g:I#BM"020b)gYH؞:c`r\1Jꥁ82pM> K{=sgyp{ 5ԡ- Fe/-{iK^ZҲei$jJwF?-iOZqi4Z.hEb4ZUිFk̰MR6VFMU+Vum-ܦjhi4Zm6VvMuid$ItC!# ի'+CLCݟ[Z x^!F 5?a6'F 5 cxq >|ކdG9:ކ |\ckHc%Y }޿`$ՓdQ`o^QAkfDvsv~f9(M_D !KإG0( Îna!i7`K>@xOE靏Xq'_r|mH[`!Mo+尚M+BOWEpz;$`=IN/gE坯Az:[5ssf5qۧJo9L3On|)b±KH 1'%-IfSSUW8X`*(vܬ B$rوQlc%V!l2W_shs1cW[Sp*${C )i1FX':c}e=TA#sֹ㪄❫zJU=*WE?tTNqc[Ti,EGaESXC԰&zlPbzQA` n{jF4[eh2`BBthEѪ}j1-T\5p==16J;Db{Cwޭ:o[J F?nHH@v(#T$K6c="Z Nɮ|8a&qk)ZK#*q{sꙊ&0Z(%NU\yE%D%䬒)H*p%;g3&Ψ&EHg4gTQۿJ Lu)VS!Rꭥc%N%-<V,K Su%Ūsx~6*lF.'LZQV褔YrH~?MM#׋3?1kN<Cmga~9Mogv7q1MۘGJH8v{dCPbq7zu>-|nhS;{^7/N_>?W{?x3uŲw@e,n{o_}mbm9*s1>3a>+kNR_9KpOWWk^Lek+M8J FczgikC(=tm='~-g_=pTZFGMPZTG/͆ -ZI?g[efsFr6+ 'd:hAg<@w3 _kdz،7s>H|bଓ hvvNKxv/Uc)@xg 7rqOAp΁a]Fګt6 Oqj|. t6ϦyW' 2I|7ZT'!ADDD` ܚ<+ INei؄pO۫ILV x/F>-Fpt CQ  ɺBt(ͫJ`H brŗ$F߉DCDZiasٿh+9<IlgK=vbo*ѯC *)Db uj7[ߵUQTw-w"[.wM5Ep% d?.^ۂ?baN&ɵTxISE1"HXcی̎GyWO~Y#b5daL(qy y?X~QH)t>QZpk쮯8[bc"IYi5u x$1A^qN'0Pmb͙‚2A_SWal1Υ<ۋi^];}STK,gI+Q0FR*r$FG2YR~duvS-cQ b`Jkꊥ%jɾZ2Ւ=Z lQ|wPMrX`$;s! m^:ҡtC/zKyrŸe,+M:t?yRGvn:B!Gao*ӫͽI/8v;x~ %gc}xpU5<( xVL`޿*߇k e:6\WR.ya+w ue#`vf&̎e0e"L8 e:EBĠF+u,P|~V/u `8v~UPY 7'#%`qsMCq5p<  2;'C`O0va} }:MazjlH j{G6+ P?V6ev=ut~ȫ ݻJŶgufAeD-OXX@lhQBil r념ZcFke]Ӟ"`Obm]m)~#bY[\EDmqɍQVLJñVd"B4vlw cupE%T!0B0>*j%UT EaWd77 | q82t|Qг@I,@$&s/E\*N<?gh_$Չba/BOԈXjNl1 t )$)XmZ3p~@K~xW^[ubvT!d *,٥Pݴ9ԡʦy{4b1M߆·)x(_v{Fs:TF))P=ܑB.wr䶇L WvO dփkM gEUNkf7J [_*+g|9E1^qbb&]RD9b+܄c8<_!$GEXLE$&Vxǎ)JQ21@XK|->!fF'ڛSbL9c%1:*mwXDi%SJZ?[6=tQRL{Hs`Hfca0,J5 *T2E;/f&XybU#R>#Ҩ\Q-#K#lD$=&Ʊ$$ qO$S++ǚY\b|-fDF7XpT]ħ9=8&MЖ&VYR)Tp#/NaD>R9]):EjqU!#ck~Q||UzNQɀL {cJ ;ޚF( K"º.߭@"tL;%3SV|j\a*clެ_*YKq & '\yˉ2#\YqWbJsT[.g ) zTbGa#H EC#j> UPlRk@*mkSt%e6F0XJ$xcGm %keBқ *V (ܪ-ϵyB⼖I)QrO5DPLF$`j$È?$U\7R'pYG=m{TTc+_Ňj*Y?Ԭ$b,! qo `K ""PP,eV$M؟kV-ϵĩyյ]Tkq]t=}L! $ -b3%I,( 1وjCvTdPFbᓱ뽓wjMN'}Y+|,. ,"xbY|ӚȚ],_! [T]H,o!;Lw }E69!%!8>;jl!"9Ehw4 ?+Ͻ#ܼ遴!?򬿘],}Ѿ~w8b5j`X>) &4H+azuhI6*Q;l<ˉ-fe}Ov.u?O_=+µVYnjcV@];D)x|be\"췫RkJZG@qvMLR2rdWaͅ0)*ˍ %Di꺶;T#P1Z>nM8P8oh9aߊ m΁h:s`>OO\HIp ϼ`Q\\lY ￧x5 3N^U#h2:_yƂ-1SPVs%Yb%Ybo#¼/YlW&Q=i ?aBJgj@.M⸆m1-hy~i]*9\!^ǭ5ڼH*nǿf-XS@5܅ʓRBu,Ѹ$JtGԍ>$|Cv޲4H/L”orsc?@pU_*Ii+_v#s> SB'ԟgy"k^#֒s,ΟIQ e[yv]ze ` S.zM0gЖhDPͰjd7ə?íɯćW瞧uT\kb2b|ѽUį̷npO_~I}DIa[g$'3nU&{Q?ǢR{27dFx9.OwhI4/mj>VfO(B莗[(L1VAPq dIx<1}@sK/+q#GEtxaWmF_/;B*:ʒ.{1}ԕ)eJ̃kv*/d0n~R٧D @(.v'$%h(n'Ik:ZW ®sAiJRct6 82!eԽJ>oou@] q+-!B0%o{NSi*ZV/P{ .2)6C^ZL'YJ9j;aP ˜բdZMړX8 cYlr'YHyc0WX (J '~ݟXg("NK^ŋDp-YOgq<\eʃfckC9)]`EB4υS5$,c )N=^ J} oN ,7u8e))KJkV$NkT+3_rvnKmKlKďw/ P z ?ȇ= ^k"@8PPn,ƾG' $y;T`ŌH%8O X3.njBR]/ZO7p\~vo3}T/,4/5Z88FTH6+I!0L#Xz% : D=XkqʍntⵔI (s)Dh2In9ՃZϑt֞%u}S|)wDUz)|X/ )'Ŝxyf2= )Aܤ o73)gK4 ,O&Ⳗ-ɲ3.t @_BN˔a_>z#IظK0sHm; |:Ѐ{P%m(zS 1"]7C۔:z,6CcZf~a(GV}Ė1"D1bsuMSr 0ث c, iU&A+ hې[qcldu eGtd sNźl`ye8$[Zּ5QP4m{F6rE:D(oq\=ZH9"j Kl*,*4 抮6@5H=xiʍޞϽxh\9m~M;@e.5pЉm?\'   QhVsp5OVgqB0ݫ/Sa͜%dm 1%)fqsLF̘zR4(2*,%BR3y3H_ƫ7zY]X@ {RiɳaBQ/|j{הxwʎ)wwc=ͫw'v+q7B ZF;3!jL _c5ZM5Ʌ\ҳ,0qUow-ޟ/FMۄ7jA&LqTQmd(O\[oLxyL9zg[szvk$S<<¡,iHq0O @Ƌq0u{+2)I£^w חx$U$De8͉,JYj\tӔH[Ȭ5osz_ܗDJ`46Y{krwW⤸ԭpBQu+@9gO?5 CZOGi(>DdG9?[ L#|OHeH V޵D6oxYgiI LC8',y+46oN<)Y[ܞ"1VܥE6qѺt:T| QR_w f$@$qMm2G3M+XCPz`=5\j݁V<ïǴL1 'SR,+r\o _c<!'H7Wj_НtTv`%w0#o~67UtڔgX[ET`qRQ ȲPgPЖpƸ(0M1b.{ jQL}MIMj'?C63uM?y{2%e[; Jg8)]O OmPԚ*Rx-Lk,_4Y͖4~6Tӵ#O6]1*z饖תiϿ>xNMYh)0~_gwnnЉ0Uv{wp1?%Kl|c=Y_ ?^O(Նon8q6;Wxܛ0N'NC%Ky7-킉6{W^Wĝ0񫯒Be;kodŋ%ܟqekP:ʅ\.GWJ(3\^p]Mt?y QBb͜NҕU6況I!֯toܥVG_WGL'[%Gg}t[$hlA:h}aÖÆH!dlx40AKcs])iu=#0C}8/FNV7 & %'W,!| *}1")D !%m y Rpa1_!ɛoFh x=>SbTu B~=Q3{dy|JsWrɰ)G9ULN&%q>Aj))J(^yPw L ԶumZDQEoNc۷N,!!84&Ya%9H!2B3D8Zet3cmJRM3 G3GN0Ck%C<Ņ{0$pg$TyԘ2) 7:p3kNVSpKGCR0Z#9,ޟ?2䶢XDID b! :l ?ꍜ4XnGN-LYi0D 3DI}hrBFT wQ( Bm$OcLeYCLʽ)j=Sp3趶Ip)"%3'ô(xZ@k$J0H8ci*qȎXmWLYP%JVbD4(Nqe#L %\%1P<')&[6tn6#. Om)JZ/ϝҙnS@P?vnOۃP; ҙP YRagulK} rhc]ռYA7I36($~<ӌ傻Ki\;yx}4|KLyTۻ2cXP^S7Y5O\)7,Wǥbs7倮IVXZV\#wF$+$qF1\>_4dpXaGv=WR;yG m}mMchus]ӧwsl4Z uMP?N^d8Hʹn%e W$+[RH}ޗ a,&RRTT:_g}ɵcV}!qjDw\{D>ׇzDw{0.3|u֎b&W14IwxFO~^r0fD΅I5+\j8|9Fį2HS0q*&\\b48ewqK^?"%^*Gg{WFA< $=FgzeښxkI,s(viKf$ەIHl$ϲ1r+;rUcوX+/XPYή$.C4"$!gZC$Xh A'RF`^~gZVIY0M ^}JBʣ40QkG0P*)ђgQ`CLb1\ uayֆK%ަsͳB k?uˮ:po{5qVk?ĉ<ܖ]#cWJYNO 7Jxh&?<\-t&Q܀`7F)c:ތsş6 k%Lrrgk_*$%@>-IKgU_0נlчT{٫յ+9-tBj"0{k"*l喕[]ޯ>}Q(BZ:2[$^JN,9V,ԩhnσl$5]3_ڶW>R}k"Lײ,tGUiu1UJrM^2ePZ)?|<hAV8nu$i*hFKtNߙ+r ;x8wQ 6ERLt]-HZvF](U|f2Y6 i"tZԐCۋR)(F\86 墂1ѫ*GeW3Eih1(W1(\ uI\3ri;jܣ Dr Qu +)_KmH!5J*A?d[+Eڿ k`V,z\檌N"ɕ~rma_{nn[Z3VWOw~{;_)}~TdwHFzq!TVVd ?5o1Cu.O?ߜc?j'OgdcӟSK~ູኘJ\yEn?󞌫Ԅs_f*֐1 -r2=SIfe<#\f<5 ~mgoN~;}Z.Y Ff.&O(_fԆ3 qT>o]NɅg(8*xquY~o .-b{m}y>_|zۮ94XX֍g Zc[gx y &[x|aǟ<383V)ϟo &hWPo`mj'~~JV`P?u3hBpK&ӷ-3 b.C䕱^U>Yk`(6qp&ì V:2XRnD/_+7ܗ!(N]ߕYyRel;d tUĠW^ `Sj[42zF{mԄt[m QԤ2ן 6L~,`hì>T Ħ讶>) 6mB"*d+nlPM'H&D5Kt' 5x|SӚ$|}Ͷ e" 8x@ia~7Bp^# b*)ktMf v6ڴRʹث{E%Ofz{ ELwցn*h p( BhFM$ ^1Bg\+qJ.z0$|RsU F[QqDI|["s54jh0a] -f^$J'D @8|0* 9PT%K "R 3: 1X̢_ukV:1k,R h1-N@b.qh c( E&ro ԃ\v7fm9s%l$9[ \ QI҃LE)Z΅ӌ܇6D kDmA s=-C.pTh2=Y|YI:)Y߮';f:g<}1m{wF!YGE׿J L>Lsù 俏&88ݝ|>B5s"hEѧG-+De fR,n](] >?=9?&LGnv҄wyj:L%7ϛd5o/7Ip_|b~XiCf[Ao5qbRL.CH|[ cPvE~ ol)px1B]IwtOe[~Kċk#ێqwn0`*ŽdP8a;oRXWg6TyQ R:NȮS`h!Դm]!RKv wOCKhhuj@µ㔾x "#Lۺ?+X m =&ȃk`a"h=9Ffy <@wo50ka|ú櫧?rqsЖZ2E<TVQќsZoX:PQcK5Co%[/&VI~)5WoHO~ 'XN h7(^JIXEPJA1 $}hRG"**)ľ4mArf͉2vJJUDޫrI|`oEE^B[15 Hr_ l]!QJT7ISr]3)Jh4~WrDg(`SgVL6f|\4 hd9h$OɀBP+9!K'I+~9#X0Hb `D[2$9Wd%u[lS>'GbVCƛH9= JJm&|&ej\S ;*peVa#k))==3J`Sr<5_7Wyu'SJh l=g - xg TڡXPgX֘KTc;aGBaz L$RE Q  +N1#K0ETY3WI0)$gTT`[lꔊ*]\|oR*$tM7k)MhKYR8rȧ<өpAQ$=R##La0PX%+m_dO=a^̇$lEC'g KtT ȁP䠪.PcP+A"]r_!!`c9H$☂`GXuXC=SQHUbmQKr Nsg^.ށQz,;0O;y-y[",90IǴRSEA %AsUgۂ: g*co~~vyk!ln9a7&Hh49b2rvZsyXV!]~7]cϳ~?#$"Woԯ~p>=Hۇ/27WbI;m@i%XS( jcЌv2W ZcF{Tч5& _*u|ޖ>_܎~_գBbajQ',IgM 55G &E!dOQӵw$s^%10xyM2H.&5u,vFS:Cj?L1׌~oYvDZnsv&>:|?}$xPnΤ2ܤK N΄D8ndo`T,6lp#Hn_3Ai!g Q=YCd DJfȃړc=w\+fZH}"+B lwLgщaiE1Ba9t0Ot&ˤP f*Ŝ.,_a_,MMj#@w(8Dr)D95^IL0! M]Ϭ$D_U+Fhb!JL6ʛ?}KKezb}Ee\ǩKM Ao+fj7@NҰ0H.a ƭYK%' )r*FLraMFQͨ砰# PqRt6 \+zu6 ekE4,hQ |~.>L. 6S++EC`1ׅɲ~޵.f$4' Z64*?">B%m r"ekNZ{CAh;6Wh˽`AϜ9v-="'[Mam/B1j M-}9zll 7Js-[Eih.=|KBJQC'B"ibA1ǣ:P'AT` BX)Iy ?֩gJ Lj.f8[|QxMQhOJ/2ǥ"fBx OZ}O\ 4Cha҇ǫB@3^׵!>[4Xz 'G<nN/<E5 ~>a,;K у65Hs7~u{o+)YRh]mU4|FLVdi:S]Ҷ_s"gkݽw蛒ҝX܍FUbbq۝vU%ȶ2Z7ީ"8!v.5аugk6=qȏLRݺNMCعE.ڣ38/|Ǒ^kBQ k;_H.j_&| _ dI˄ooH.!:d]÷+6l[R_o7' r)HDf3'qb ,c"=8~V,IN22ńJ9өUyC$gRäSce)!> *yGE">M]Z5ZrRȰCZJz텧<)mD9JpaJn5,I5F?Mng>AJ .I /qH%49$\qgc=9lnm2am/ nDBzn|d(D}Z*|M:e*_#$wCgh#QZ\nBnN.G!F7~7&5!=yfT( !2p_A['>u#@;:[EMG9z\5nWLS&{UyXN@=? "4v/;#3qqxѢ$1WEɝ^>7|g-rϡtC7JB ECk˭ {ҥX.DA4n>/u7PChs*npJMynԘsKi>=)0gd~;~&{яa׽PϏ}yBLuRkR_QíeAM>Uxt;|1勌+~/"ȧʨb$ O>:*ª0>{+in uȰ~5!]T2L.Ms0hnWP,nsRbSnaet-mw mV6fhb|4w)(qKm)Q˒7T2O'íȜcnrzljOS;!.#n+`[%m5 QVVkQ^JlEXRl뤨,-5̆aLmuʇ`$RCԡ"a~t$gd.icK z'/v=hu{bfeb?_Unn=n4hNmY4X; ^xvps8~ /q--O/]~Ӯ^|*: 6{,q# wYb^|[WNBBF4O@)s:'GߞRs|?Tz&A+.QKl!gDQrnsP!'^c"ErݴĪZG~KoԾT\ (SU44W.J\Ή˵ݏK3 c7"liC8~䘞28❪*gߌUWy;-Mt(-?⃑8*i9CΌ^$z"9a;N_E21* H^'~={5ߋ%!xjqe񪗸bT+_QGO+H~8xUxn\&#C ?v][oǒ+Sŀ^F@E"$e8!%.l_ TueeZRRXZ\6߲j'm6c{E.e',}2{ɗ]rZm}^rGlcx]8wv #G0}n٪qa : ݹG _.|[yp ̎v C|SX̆e:YyvS@ib=Wp|Ӯ@݂oFwԿKȳE Ud7v:vMON 4v ۑgSno"#LynѭrvttQpܴC:Cv9]!Γ(}S9G#OeW;/jw;gd<T֭mz5~ G ,8Jxv(E 'v FmuĶRzw}3il϶L(E%~垠GǽEm ̄2ؔud) cS-{14̀Cg~|iw[ݔiVݎ! 'ؤg@&NOISpܞң2Y7`Du/c7fn)|p: YZ/*QfЫzɖ7<. xSf4W?E.|1~N"#3?7F@Ƌjq1bL=h$#F9i[+W_;&Gp Hm Jk)zFݵWî'&oo¥f3f5ۈV)$\O( j ʉ#"X1&a?o@NHE_3{zzv\Wlrp|[3p0i-\B{,')T[aـi0ڊH؂ S`('uJ ;u];{#nN})97 kgisO/*)@|¸9My5X9W z00HiX7#"IYI-u Y'q19~b@5]pI~ܱs&:#ꍴ I(Vʜpe82S_Ϛ,kq4w=Iyc=R%_,[,&]F}TC@.:Q W`$DqX[{kM LeAYopcHcƈ1\ p-A(}X:[˾ `m5 h _$Q!A\( , &` 5Ho{JX`[ ADWa HI0+ X#I# 5ho*+: vF JxpՠPNik\ DZEo;eUnANbY+Rq5nw~5`Ou0c/;ɓ?2$braʉNI\),! Vʼb m3J)I 9ïvYԝ卙9 Und~Q0X*,ySwMQphׁ`Z8= h\0XX |~`bł{g١`zC?\Xt~`b!0_ŽzmŲOapzQB#9BQ )V=03"&4i?X&~_,Œ_,{] K/%?\ޞbk~_XCr]Kll{5;&K7wr\țִD$ߡi__ZAjQ!Iݞj\]p{=Z#e3lpk)Vg.TSMkn&׹Ts -uZ0vTg;@S-P[6u]Zc:ؚj>}9#~SJzk(gt9N<\w'Z7C_lM08:[n|X8xfW?Ζi7eS6͚U_n(Uo>cEyts? x|1~dhW`'̯VZòJڑ,/ºbDtruo]xά[~ukCB\DdJ~`En43[xJ)ڝl껵 !.Y2%苏=n:[.MD'XSvg+ Z6$E4K0{M2=Z\ N36ncI[~-QCֆfԶYX7I{qL 52Qܭmi>mݲZ,R/uELs1h$:ȺV<h-ukAB\DdM=<}떋A}&֭:SEwf+Z6$E4Kam린/bDtru^xJ*B4nmHȁhL_к1221hg>ciui+$E4K(߷@lԵ dJ!o{CR]j)ݷ@5U˖zV5m[Qgm%3GH_IK[9a>" 3ؿ3;<"̷wcY Yr%ܪ&us1 T1 KՄzd~orB3Vr%ܪ&h!/ǜ2-%\rin(S)V¤ܜ}M.¢ $$=Wpp SIEAG2:>b^xnaM`JQ,O_x2GMMp~BZ2s]ue0Wo ~</a@n qxBGLp3HđH[&s3$ByBT0NS% *RhlA{0 8"έ6a H ʀ@Zd̈́amf7 /a$q^ds8he:J7S cx^@-?ͩL#^]ۧsv ؿ5Kog#LV`]IA8 ZF@dCI++jlc&p{HGr2> ŃHpPRjpl\aEH-O/r1;GTDsQf`̘ cT*X&B.CEd(S0k9((2:8YaAC ΄vB走"fgTk%{tJLj@$F"rl}|B!2]?!h~L`}2OSgb{րv5GhTo.Nz} Rj8j}{^=Av,gȏx9#C[nY6bUуV2M8dOVH$q؎L xDbXLǯ&G&]'pf dBlMnq>\MW  t}مײF FoU6 9T3:7S(uG,ǩBQ S" OI"k(A0Χ6Onpz1@`>L#K aҫ|yvCn,$i5wYn$Z,o.R;l|TO c AA(V Vidb?dӝ`p~-ߪ@h 1G2'䞤!Nр'ǩc㈠h3 [!-Oqzct]$!$]OVJ;Y+o?~2gRBNS wpfhkG:rw`=\:nSWU>Yq.tށ1%_Mٯ! nf[g9JT<R-y&_M\{ubN7[¢eמ;bB::ʯes]91;%8xV1bK\} ׳U6l+iC3{}w5Yo %]Iv5 n .lZ5_M'^tϳ8~g0ѢcjC"۽}Jiߖ3>zGqv$OvYD34|*ɥXz` i KԥH-;V, 3^Yr0?]Jqyؽ_@ ($<98eҁ. j>j: DyF%1%i:#2p2]1/8rc1S-GG BqjF瀓!pc,Mw2-GTpm.uq+Zz8̙jr9:H.wÓ+0;r$ H؈D2/`["D邵J2BwTEMlLrr?I =njKe^iÆ̑ mh W":vm~z׺Z֭T9SF6`3kАE:ulcĹkDq@V*)]u23kА[:7S 2%Unn. "o`O6j Kԭ|w,HڋgIlmoY,f?_]neba$va$ ,쁤wcL(VI;LZ _#C9ۃ#0+dzOBxf^xN)b53"` twon4-?wA`NAm3+XV:pݹ+!AR{)Y- &1ֳ+ TaVUqHV+ܯ񩼲PEj9Ԭ*/G~d9id!~ `ÝkZ_j*LL9}uS-T;_eګmZ9ioYp>9QԲ#Oޤ_$: Īf,owAn% =NhS3|'vٝLN?$+.6JPryM{3M)FkHNSDyV"PE) mG#Owi~KC\meKC0P5={۽CV*5CBp#G}rYxVDsӫi2pس33'u3v:dŬwu>׊sacIJUv)T'|lQ̮/ixʱ>(完9'8_<'QEB!h[2'#dEژfcLhgN{ aĢ"P"/bji4iݺU]^8ew\جH!t_g %kTQmNk*R 4R=VgzNژ5:XwR3%C[\Tң⊤Hsg*\$ТfjP֜şlHNnﰙF.7> zv\Jx$'$ː$TgKPdT-xr̉w'mh==&Vw'`r1/K!4<H&Wd*(KZ]ĒG˜a>FsA3fD鄈9@*z"PR`7ó$}x(4_w7+`bQ0&`SVQDž0e ?F^`eGgeC0О֎=Q Y>!.8D5uLs)G'- ek&_|[dezU}h sw[i3 142ZAՅT;aXHA&BVT"|0Bxmj/ʁ'^TqS)oBŔ@butՁ5s P|. H'pjuYnMM:mZ:|$'S\$#w? d+BOoOVAǞURe|5WJWM̪9)֡yiDP9D#ʌ(53@E.5M!i3]O"rO9=9^ӳ;hkOD)ÌleVÃX>aXpK3yOS9k}SFg!J ->|PyIQNU~dA!0nbcf2$ QceTb@:HF"`ׅneIbQ"iٌ1Hz5H”Q2Z!p„Ny ri""ҫe(U"lI}A<]Lu:Mpf NZ0콖J$ש|^W\y]<:0J_nr9iE9(zcHI\8_*SC(09{ƫ5?D(ru3~ bG=C\<"["/arTt @:R__/+gn^FVtN㧋0O-.I)>6">E? I c.ՕUAUL`TY|s^1Ձ5>ݨJv"> p/gF`ϯ+^ϒUJXRI)<Ϣ"Da._0BsTtKҌ1ѧtsQD8S0K"^(b܆b,2I Euks^vkƘ*b\) ySS<9\k^:#tJSo)j-Ji1}M$8\q#HŤ\V10Z8\9$t/ 48h瞐]#+訑r-"R(ѣԋ[ ϜBBf]= 9*0O\ +0O[025Oaq0rBDcOeϿZ{NFH8l;E}AAh~SLI̺AHcbnwjVwVְes<;NXM1 'z<;x#pHo>(ՇHkёJh 7y~ %Ձ5PLk&ggM` 3Uh45LU/P L2/YGK̈鳜͈@T=6~(r~^ , %]UG0HN])8fGmf X(O>k0BOSW,PNw1VXmةCб <[\WBoo&ՀR1iRЛIZtAPoIXO՚ly>ŏWv0Q#o، :jT,g)ֳzXRY%`δvAtDlrSᗤX0D 1QekGBEQPhNG;5`G2$50x1-fGg%]K]mo#7+`>%$In`Yv$ygWZRkv7[$3mXOUHVfH&6Snpщ\q8VC9 Aڱ@3f1(>NMl3(9E}Fĝ8'c$3\Ny N\6 틙V" GM)R$RL5SJcm8ӕҊJQ04xau"TRRJVl8mXi!o8y+lNOh _ޅ]gBඦ* ^ ъ`c\>|B:C2[#=/]D_ y݉v)?DKڐTfVv+¥љ4@8Г.=R4.^w}~)Ty&.LS2":I@?qNX?%ϖe$qvIE$> IUG0:[Qp8R)Im# JS$ DRvj|4%lF]3_OuA)/ VMJB+Jzpڗ`̪h&+[WY1 u/rZPsyuZ\088^'u~,#׷/-ߥF8NNq]JdcjʞYz?)3Ts.?Ow;8|`faMpk`?|Z4x}h>uXO6{O9M'_,v'~w:e~qB/sr; I%Hіhzx4;<}x565D+@ȗ*Pn<7>ܟІs@_G2p8!F~?ݘ;s2Ybᓗѻ˷t0'z%^Id|J "N͆7x'sDa,\YR9b J߉`FVVHBI*]UG+2B:'^#۬LI0](b_i**(Y%xUZB($e\YG 24JNxpRUXK).+)eI#?%\ ෉e(Qe'. Jٽ*ʠ(@;-z Ji)PJaFzfhU)x768BX.&?_!fʼڼ|wmV@rٛxˀEј ͈B /c&cQ״U&IC CJ.-\===I_fP:IʨqN` dU: \iw!x$˯y-PYV,lۻxv\6^$.I|'䇟A/N8{>%JqO/(۬4? ~]$#Lc!_Tj `Ǧ߉oq` ۓA[>E?Vv1._łNbwF*-ҋqXPqYjw[ Mn*<c4y\s6 M1AaL ִJi<jo7tqu\\0DYﰚДQ_CcH'kY4 hDM aie%̼h wZk#kƤi.ħLDvaeYbh vF(G rG#69Bħjd}H&ud0tɤL N2%6 &zhy`F!irXdvچ1c1 v}W0e`J^btCgK^2+MBE)4Q<ӽ*#mXooo:*6)BEր9=ŘxfQic;tnaoS0Np6.N3k4Wfw$M8rPh`@Ƈ=?Qųqw y:|?YOm8 /]حkf3_:\kvc?*!'wޏkxc1ukyd~Kљ>mhݍF}OM/u0i@b LҴ ~ P;dj}CNkǢegY7V].cX΋U|҅t՟4?cNTkbΣ<$$FA}^ދtS2[-gT24wÏUG5<~Ÿfzj&Ǝ$cɌ?zzAQxLx b|9hqx' ܬJ+qOL#B乁ۙ,D]KXuOvc"{ ˊb灤p[__?>IB>y U4Mp0zL)ΛU$ 4Y`l灡'nPEQĩZ)eE]۬V~&3T9'KyV;;:1t6Y@J);"m8<4d>x Ϧ=y#=>K}aW7zY^GWr.T>{ O{Ѝ{\[(ҢqC &f((ǻˡ/6t ӥCp)s_~6CH& PWu1|(s6<0`k( pSCۍ@Q2Xi-:4=7dދy?B $4=zn FIV,6 %Dn{6;bdj}3a4C1/0:=C{rc O1p ^ pB]Ӌ^}oY5ZL2^ xW%ٵn&e@I@4UU҉5IӲZeWϷKRJ Ϋ(,#WR&HSYP([/,q2JEOlMUc\o,,+xs0J~{kw(eU䃤%1,Fmdւj֩{芇2OK?{ަgdQ[&b(kJڸw5 Nr݅tALowѝsn;KP9$bysLA럥Z(ۣJq Lj+*@]{ּYkF=ժͧE'vicOuuOgV0(ǕUdeRV1{8 /mkG8з&v6b@NݗzEYm=PJ󿓿Z6DoLOL6.NW9]1(m9QE2e˺? P9ct\`Xx4HZBB+\R%++Qy]9g\͇ Dߓ~rY~٤DO>lRAz;4@~6"z3d}j~&3yWKA$kҾM)yخL f)Y+Dvl&ҁ(!hI'`ѰZoULRƁ.lJժֱw7uSCt드 ͖l7^/PɄ=\>D;)Z̀]70KRK(eٰiL) ]qJLr hlss/EUЊ^CK\(4S<8\+(u?ǃGx>(J:\zp+$P<)*+[o2, %J(D --vo 1FLv nyo ~gq nyZdÙQ:AmǨ?AVX7] '~ɴԳSC'gL=:TȎ#ey=L*b1 c&0W%(A.I(s׉:%N.ߑET"jJHUDs)+)W\N䤞jbRb&Q^+Y|þŔ_YIHe>} )'gw`AG؜~ƣ!E)0Ԫ C,l邑RxT|Xj>XtI>X5Ă̹?ϫ;07IDL Z㜿dhԧD+8|ΨG5`2tF DjݜXosN@)Dx;ؔhFk%:zݿBv")ѻS$cƃHZ .@TbLJ/*m58{7Z+qA3Žz2YƥYxmK_R*sr4Ё P1 !~gԁO\>sv xo?taSgY3f4)Pޱ>>)%Qn Zw6ׂpAJ:GY+q{4_HYQ-uE|ߔvy- M HV59]H ePb8#9J M8 ؜i.V8znQBr YEeDP4/a80c8l-kmʼ_-5F( xuX5{/n* tkaZ6utח؁^ y r*u~![H) d\@inn"ܬ;HIT]TAI:t/< ÿ1}ejx Pō3wo.W=ÿ%#E"WWoU؇$oL9t(7e'?()皠2 ~!bCgUdQEN-M{©~7_oGX-q<͂P9Q>AIGGfw>{Ńnx'ZÕ+ϓ{M5}Y#J9Y^Q2c8#^CTF?<6o<̅UfƤSIEHO@Ai<(ɀ& ?5m~;,)1{:#A7^:Ie" /-'_i)$Nvx:l3Kԁ@JS"'$J \U( pъyc] fY+N*ܫ$2|w1bլVQw}uiET_.]E"ŒDd%+!@J+E5i@_^Sz+wm_A3ӫr$%&qE?yKSb"S*$&>* 槈iBrG(#`RD#Z@=vg4 /Py4>5((ZDBH 2P'\(syl!8z:+ʪ7>a1. &ktHF= [Mn{ ?8EkD0k6R!"ͬ 0-T8=ȚKBEŬNϩ"Jlȝ( 3|sp91}fe aI`vJ[%cD1 k"X2X !3IQc*pӒMS %Oe5+Č40FG0+aYփ!~?ڵ^.2& $/W #8z!pPwy_Lc - = n=`1nalpMfp練 y_t zɟtBpu[<>ةI7xs[@ xӅΖS|^p(d6Nu==xͰ'|4|+z!ǜ\]< pnBolao2,gt?dsr1hYb;lnX92?/ޚ6<\e,s>RҊ#b ɲ,㧳 bwUuX,o!o+Y΄[ҚlVr_6 = o >V\,xJ+݀n"ܗ~uZb.~kI֭O:O_j#Θ]V2:vM\n!G*Dm{nޔbD`ƯE<^{7 u_猻v~_a|_;/{?ZR%]Vbu76fIpҜ_qgIwqʝøCkP9Xѐm lHs4bT)U-G5(8^ӧf{2);ImӍE0/3SfU]'?mFZlEjnLz2Q2!^̪@U$!߸,S+ނCh'XQyO]!Yb~77 ?\13 zb<;t~4hO6>JX[v:"/05ݿ'FE<D+t䇏d^ǰ=)/>d(o e>qE_CS꧲=Ϯxb1\^b"&1GGsp Dk  fq978Ս,FW%Q. XQK4@$Qz4z~ !Ib!;' \BeWlZ`Υ>'w,S1<ሪR'jGL{6ٛ6Uh>Jl\*uC塴2$kQSӚTǩa|eHMfnS R\`IͅXTRKLA WOW~}XeaўOcTD :*H&֦7/hC_n_{ψ cW`k8QL׌I[ 4ӯ<p &EFm<ʁaqMW"o:PsiO?쏭 B p޲;خBEZJȱn%ꎬ!G]E:0*-׻gQv3;AMvV3ꐓX JtQG(Tuُn %]ń4;Esu;U$r>Dsr<ްĺTj½ΧKa&:*G+QK :PmujZl3_pɨ㪵!, cl}K+-χr$jSveWDhhz6kU8>1UI%"3J/_kþ$*Hb9 4.xYQK%рP84 H3'1= ƁY,6ˆBZ2ۄ40F rSӛ` /Zf?7^K{}I/פbd߻Jq ڍwӏ7W/߱Qz i~7\7(  U!Ł/lt:|?Ɠ?&y,'kDoCIZ.>2/=L2vaZ.u =05jRjߴ07,m~Hg{EJ͡2}o#]͙nz~VRsa&P|?0.-a8L[qFKud8#XfO+-cnL[Cl#~-l%}追 ٍYsֶC~].YifX-r8oa|_`;/{?Zr+TֽЫ kpnS+|-d@Jչ;Ө,رlיzN !O%5kZfLAT55JĵJUwP]R2YR2Q&[f!#P6cF:h2{Qբl<6aQmYﻉ^X5GKK>qʞil%s4k8qY+=>G3!wCG~_"[ 54/n"k[݆-D$'{,FG57q0#bm\F6^yP3=~SSkr&FUp caOkpR5\6Y_쇾Ԍ?DM*ht VlpՕ lhS,[SYvD]rhtɜ,5i{.eU'\%9dFJ8/C%sN=T.YArhnM)'e''\&_-4g+ƒ.azW2/wKҋ'ܮNP;5\\4\R, @ (Org?+QEEdT-~|TW`e|YQ,DnYV,n(אa(c˹&OުfbʪlfEY37|2Y|oad9G/r.si|l &ʾ#Lh`My5Э1x)rf"}̔YUgK8 $LV?ċLo\DO)*Dt Jb r 8Oo?ڵ%dkazFkSo2>6ns5K;ʯ^*j%z)})W@9 TݏknΥ|!<.b##g操Wէ*ϗ78RCr@";ܔ>{ M {=HN *W٢2to\h>|.~2UPB,(.Gu~t( E'N1LihrJ6zn[|yO`>~IhZ0B(8mdFFcIesOp[kٱ>B)->jP~m;.Dc$D\!ďc**fTyoɪԥ @)(mBwNŲ`gS,{"lfvRg $n0OZ:W/R`TVPָlB\bOôS/tB ]4{+htj\Δ> duGu ܱXV,Xo3E!ikKoj5ƪ~p) ukE hvebQ}=Z @츺췷^7V'nnzinz8sVw 9]!mR^J}pLĩv"%݆h%(iL}*xqNg+ƇcGZԝ~bOEQv7q?D*cN7SC*Ѣ9B;smj+˱n\ :jt:W=J;ѡhb bg(1ۊ.pL*TաM/(uSK@ ,_ʵvBU1I;4-VZX ݦ,SIBM {bwٞK`R2HIj+!I$ .PB#W2$C߮\ *s,Ea$Xh &$1dqgO|a_>HTT,5?8W 8-40F rSwm#Icv!a'38lER67('#AI,YT3A"Q-uuK-򸱞ƚ;S%We5"zQ?.C]i"\-X"u,`ZH.^Yuʢu,xn'|-;_XYTm_+105A6y`@["[ G^VUw ;+UW[lx7~95UI*\e,ZDžM {3_} ox} L0jYJQ&0MH͍G+7`>9.]gm&?vU(( 9&لi#!d"bLOuƇzuR݂6+;8o% 2F7FFc`V, FJbRo[+YLpTh%uR]w:YY3tOmÃO='%Z ΍؋TQ@ګ5@{;@ fX}eG=Mfe1!lY{FA":upq3NKp.XR:%YpTÆ`LQ 6 !1I SlqsLF3~D0hLJmM~_FXyFp+Q 2հ 2EA#>a+!\? &@Ձnn];x WJFԀJ#!Z֖˚K6j^XtM0=6ul M& EmbZ#n2&).cSWQ4[5?+m"w _K00,i J@$%Wʚ]YYSz4-8O|jR쬒ԚSQ)KY%?OC˚S=ՖJ* |'JRVz7+9hM Swpm1ZKDŽz}%,{dWϝX&ZܯFN~Oϲi'K:ēPToAu VC>*"R_cqN5$Y?0$ij6{-١% M*$EXɀO'ڇ0~-p|ᦦ/yZ/\ 僢}(ŀ,nM#>Lǔ%\%%Vn\aqՁR0ųHI4 ҈O(QZ})(Jʑ/S?dP)HFJ;L~u5Ў3ҎYF-r~aBZ;ha4Ŵ4d_bGՋ?WTͭ1o:O?k-Ѷ|Hŧ&IԉaY<'|sq"}2t8 6@B1_/^ͦbʕv}<)dXzcKB Ro>O$pr4g@sdK,Ey.ч}ba죑=gu6rB7]82Gu@E=sK7bN3Sdzg3"g2$/\`L㓱xR歴lE `^{7R#jg SQ*[A3';H`J6 B 0X*麲d0xLVF7 njOcgb'%dd>A.e}2І-t? yV|=KTQ39Uu 3D FgeBԇ %[\aHp4@@n;TuX Ύd y.uNE%s(gDLHHT4spgAwgjF\+eb0sܴL0`p}W'Xv CAWGN&kp v=>! -}&#b|_C4Y%2_fb"e1Pr* T얌%Dn}x8Bk`VJ*-D *%Է@JÍR[MU.BLtJ/ټ)t^yBo%’Q궶*El:t"F9!@Gc BJGl@v.2@!8SӱK9Ƣv0< R0x&r\Xg<,Xe)Ŕ6'r <'=}GT/zkL}>dE?R߻pbP>"K='=TCLagѭ `Bsz -ȬW5,@3ds X8)lii)JϢ7oGm6:T?'Qɀ2LDmf jjOag`*olkx/G 3ZŻekQRý=`os#j]s~UuUZ}Piz/` YeT/ _}\jd0ekOn6.RB7m2l oz`wL G[/%1vыk*4NAM29!y\C$"RH7%MhǍa9ǜ*_ގ}l)dءIlG2Fy|Xy6>U l%MT7MFcW[t3S[gGbkW 8Hu}=3k9rhou6P0.٭?VoaQOS?woYB+k޴ബZCtRj Fu݂6GXtF(([gg kQҌ`C3iUCX7Ƶ AFKp.XR2&6Y }FG3223qڊֵC&.Ĵ}:SZ@Q`Ymd:&[}>|&ɵQ۸I,>ռQ3Լ&f[hĭ{J4o%hTdT}@i̖m/[Efޢ&vy?[N[ 5#cW)r \O!8kK"y .G_ 6ƘĚz+eJbOKdŻY1P!;/;(/;U ª:w*eˑv*eafFtURRj͗W۾Bb5/+J^T`V0f>[lTDC0t4Q(v%VYFVp7d+e#lt {_]<[U`e -8c͓ZO*eI ;?ώq6iJZof\|AqchBu!40DcX=]۷ S{-LVw@MRKzE$\ǢS14/rY(>;O)7"FtߤSp"?e:bC4 dvu"9#kaRlCD]{F21k QHq3ZĦhGvl@ijHuDŲOY.ir 88y=\Lߖ,/ěFz6J~ '\[|LO/O8W'+[,ǜ2g"y*(wOoXvͅZJYB[eRϽ~ޙ; }SR-gß"oz{E}{ Ͳg WҖ'VytO{\)uxYϺ؉)2ϣu|k!~,kayc뗵Og8=Ғs'/#ݷ?}8a\Ƨ\>0T x K9; ۢcK]& gK#uR4M ,_nLx;QDG>:LN#n~c.@8z%M}]e]+rc)k;S.~˙,lj=4mƎ٤jQ&'y@\'X_M(Sib rByRDr]AX_XS]#|#&1e0պmWւMO~'.,G# 1-h l{QC _pEߘ!&_T4 U<] *B_yh$R쇣&N'Dj7˗/uEw-Rdpi\cl9fy`;r ]7B%= M/#}n ,?+[ ?S5~~~@.~p9LnZ{݇ނy}TcqN5DxLv!BaFD`L#&uK2Eyͯ]f..,7 9eUNr39gMYUdQ2*4,2tr0̤F6̸X w9|B.:C]F`t9]F;Cr0lu9`*MX3`/2``F˖?&/V _"rql0а B#Em2=!3C !!u%# F#m"EkMȢrf^~abQF@\+۵H+pFA`E(6Ȕ 0ӧFn?@zmwFmwFmm*0Pm?_sɔ:' ee>{߹ql1UTn3Xѿ m<6NWeZI#ˌ!'W-؋9_&'Z_bAm  ϝal]msܸ+Srƫ!e{nɗKHuKhY_r4IPd-S$OѯV>U pNMLʧOUۻqbm=yڨ/(k0qnyq;CwZ]pF?\I̗sw&fDU}e_^qǠJLi~m^*qT}gX'I4Nz5)4fJKi2Söͫ:=z7_!s4IX%FTwM^h8NJؒ t mjP$:zq` XR"[LR%-x- <7j7Ѡسo0z_<ĺ&J7}qaWeI꽘[$/kVSlɛRqQo/!L4Pt:Pa}x:6A|2!YPlfJ!/, &YHrC0Ҕ, QR)P UBWMg < YN9˹*BHGT <#1bYм/9 ۬$ ΢2kOðC#L|vr#crX1b9F,Lj#cr'beURILj{偞^{"8d  $DTUر`GevT:ܼ^/R4HEF/EHi"^ы4zF/E"EZ@( nF/,0FB:#Rf8'av̷B`Kc3`GM}?WaGǸ[qH]7Aݫf@ՍxnWf3],}8ؙ ł{@&Hr#ͥFXޱUZ\ݯaY3~lx-3'-VO 5_n_ry^Bӌk6eInBgeIu>yk B5oZ "*{cp}!F_ò?OC :*"uj/\Օl( Gf8`x$:_-%ۯs{ Nr6^绫+*T*"%쬎ϐ '_]:1NI"ؾ3@3A(T'>gzyRGB oqgS[jxtap_d`y!po6eDӜ3N1kdz4l.>kr@50j8{|fĀSB`9}hB)D9<tOi.F"^幠ӯK/x(B.pN:| TcS4fx6$/oۋͤfe4o{-}N&8[fsnT>bx/M>ziK^|lJ 4gPgFdչ\4( Jf Bwc6M][wN>e6G(~|R0œsԝeݞ|i?ޚ%eQ [@ᶗqAKH&t!dRLa K& -%[ƒㄇP &[ 9GZ)ܡ! !I_ .yͬRX2c)uJQpj73,ELKOC¥ !0nT,yeY ]uCH@bb% gAU}nXgV.ѡW]ե_{߃a{/A#A2" aP ,-JTQ4cXN5ZJ3tK8QQ|h]b-s/ʠb !X™Hѱ2+ۑ INވ9; 2e*SPD2:$>SFTa25;u6:s+AbD̚-*!JK2i?( 1>6Q:DzLrBබ^?^Ns6;LnV/ IeZ\lI7ɹ{F6EuHu턫B}t*HnAhh0|1+GBXvrhŏJ䮗@AC쏍ccnGcE:_"qJؒՓw Ta>׊lr]ֺ4 qOupL{XXN'VjȜ\DBiaCJ)d噳D*3KZٷN/4dZrƝ}zaEu fZ#{Md0!p5wI``N`|'܉/:vOFu"S_Cq~dQBQu[89 ^]z%c9H l2Q*K;/`f%)TK)v) GAS_B|%ޱ4d/A0*Jޜ]ҔthwW(oսII VX5Ix4ҞI,4@;dM"NoPx-B{; ?-/7AzF+k=lM~}~`goKl~LAKF ?{Gq2CtMnT1Hzۭ:UªN&붴s\'f{K?PS͔lc>g9nQCiܡIUDD8-tTHLH&WS:(i11ex6?U8I /඗w1YP &iZTAS֖ba| m4 O] H?)Έlc-Zk?o6Ubw7߽Azpm̴'./ΪoLoVny-kriVX Gu?(^bpȚx1@so|iP$7=tFA*H` Hf!Տ-.&c]"10()~ 8G([|.7͙"Ԭъ]LzlGFajx'EƠ%x" Fӽ5@ݬKlGj׌t? W*68qw5I'oi.ĩKC8^@|H$ =WvꍤlV04R9F+]1u ePtª9NsLe\/%BvTNNij'$, ~V7'mQFi߈kQ9آ4 nQ!ExJahF00 uR#z]LMgJ.8䕳hOi:xsWq2̩4aoHS"h,-Z!Pu!Ex*E))\Qq0c}#Cpo$@!fD*zLL/Tp͎FnF4oδ;5va}s Fc&vH=v@pn01cHQ ;I?7C '?E+}Ă*xk,e3ߪoik,(xPⷰ 궿o'EI4K=Bm̒;z7޵?mkzo{Mf3I37mtp-=et4KeuO$hqߣ݉J>nW0z,_<-E*{pV uC૴?kj7Ge )2"s5Kt)_fcrw{OFgcP;F{qQo{rk_㉽A{+El-9; &lyzʾ8}?GOdtu);̸3|vOf|{ޝ,a<3^>_'(0:M일Wz!u{Ѳ0Guх-ui/[.\[TfM[IltHWEg2=WY,( d:G_LSFar&i~~ FhdoElc^E˃rlyQ:?.߿:%u_.$\ՍӒԑDq磃R-~}Wr؁Y^C*ʼn'1{`j0̯SRd$Ȍzy=|:2KtʶϗǔUx=iL%E3eE oN{siìVz9XNfL:?]| tTv$w#!(|C+Qg&/e?jhkR6d:k IĜw~)=8gIoܳӡ ڑ> CZh,8HC(Wtg2}2>Uud)k>gf͛}12j?  H0_L6Mˡ~ikh-v5R!kw˴2Vsݬ̔LF镹>^k,`'X7Y0^w4JU) *\zk6oY:Og\W$z6T۵leW<8jS2)5X#9(%Dʌ#!TR$R1%lcv?1ֆafPJ+wԸh:p:5;剝rC+4]/Ζsk Iw {bSFhmE"!"!"!"!"!ҪdA7xc#C<6RpSs"{b-v4kLu'vbBG"m.E Q;"G>(R"(R"(R"Er>φ(H;"meUj,P"=(RH$aQfhovu o6lf7͆5ncU2š&bcn!]\U+;kl]ʑ% & ^'{P˺Xt<}?osySWJщdTE9F0 '!hQmpGI˙萃rFlpG=M6Y|1TFxP 8#d]zoE|ݵ Zsl8snfPȍ1cPqbIa0NRTr qg9hڛ ӺhְP~#'Sna!IcmF<`אE@v` [p)c&q v`mkm+dp0g6$nD'vA]s&Qv!/ŷTN yyyyy_ˋoiU iȋHE~Dk! D0hKkCB k.%z37gd5Y#Z3MlMe+.vP&aȭ!Ds 7Sgnil|ְX6ARwbc7syj# % 7H̗.J9%Bdj ^zazADy@3%23 D5fȔn&J7Ie˵gBwϔn尙mKuV@[m%ʨxn8O(OU2 ' R`+MxEYNiD8ʄI3BlEVo֞pG FWN ~ƕ6.mo˹vx 9+O܉(kLӞ`JS1YHɀ]cvnRk_ Žyb<`6h̤; vkv;c1Ac--ƔQgӏٴp*yl M)p6Φ8gSl MkM[ZB8gFo3kt#I8 n.,0oνӇo6x lfof]o7 mJ4 cfչJ}[bלhOpuÓ ك؝ヿ{I {3Gۯ^y?4FU9nO'ݞA47fDYҷ{ -v)o>G}ã 5Z Z.g25CwS{4Z]c~I%20H~T~\󲤝&+]78#0W\C5jjɛtBat1aH=:|ПNjZ$[phЃOnLV<>cC~B`.fr?y eP%q HXmyX NazǷ!r>\mNfS`@LusMzE=  0ߵg$fT5־sLW`XwѳEX4ӪX|#ASh̰mh{,)&߉\-OZZKS҇}Q'*_"`ޝLe٪?&YkYSVdUqʇW9@8S`B=!H[v$xIaPyTd1Ɔ2#EJy^OZ? MTƉL x*OE!hjHAyBH L$Te"SH,JNQaV^|ϩ (ʮxxs&Y(ҷ~rAIVGp~%\1nOGݲ E1"quo`'3<ˍBӢĂ!*PM4%J^2<< F>^{* &Lԯ><6lYL ¯<Dr"U_y?t.ލՒBMY0FBrm *AHbO6BTkz :ޜuߜx[o1so<9.V +Ն Zb0O,/ŸᡌH$-:Z, 4͠QjlV1rXS[L:|6=Kήn-ؽXŎCaA~y"̓9{?oZϽ!_ɷ&ctg"HiE4aE7&Z" j:1'C7VׯE׷/]^T>y7#Md4yG}H~-l[Yr Nkv1+\ jB*;Es.-SD) yZ43fqbcR9enHp"MK4ɬAn1hd[Q9Q7Il\|tm.AHa >+9D'3G%UQu%cpT2 ~2v5(~IuV'ʻI/dh[Fh b:)Z⠖+ax1m#Y^f7~qRv+uvWgI36Eʼv߷("J" t }ͫa~uq/)hxG2Uq dc"/mCb듗gѼX1"Ru7)=LH4g#/ 9&$Kx*@L[LBr 3^&:ĥ`!8[{^bW2z sա SoF !]&gQp(4"1h9zyaY&>L'R\%nw!.4 "jE$3Tm5,xzYքR<2C*T aς#;֛"TJ6[2fob*`wCvH:IgL`LQYsp[ cop6 j 1Xp8|z U) 0i]f\&gVSQϮ+ЄIVh;ܧd8 :#uIGqDDTNDpÈ> :uV%'9O ?,1܀Lx&<| %їM"%m:10%x1ɜdYQ}b|'y" 'J*0O9,B]n&F^A %YʺyA(A(*Qf6 ʑ8@ Xʐc^~0S28I\*R>Hki#q Z3DK LJa%;ALFl(D(fCJi'{8bP,MlB4LR?~5F!~s1e-% g-s_$=&/tXuV$|l@%~fbO.U kPRw6Zk|mV1*x6B鲠b\Е ^n Æ耺\t_\x+!>2EaHybj$GN_n{%=gR^n\0zwcJɧ{?{ÝLLŒщq'TSqfX#&"}i}JBDJwհЄomYkߟ3͒ɫ &^,h(br}t1oC65KUljD Pq P܏a2HHh RΥ҆cɰw p&[;JS& Zş(%W зlFА?Xbp̸f[1[[ ;wXa3S=xpdX {PW9 __ ⱸ^ǂ )Qn}rjf7򔢘>hvwVڮA#{EeK| q|xGa S 3uFV\% z}nEG 5QD%}+fdO~EI$J!wVe GZP!4̈́pԻJȰD h)2DumaS}j!Ig`G6cy1 ` SIaך-[JFT+C\%Պ.?oP o-x7EFDb'6m6ށmDm$jP;T[ivmN!=uޡ=8e_!KEu^;"#(cdY_T fZTdJfij SBHՄE pIl5?Mn5,fI(})Z[֐DBdoPaէƛnb?x _l)!/c&J'9պ<81]|{B"#]GzЗZCe9yu_/zbjՅDήb[ j81=8ݢ<6yjPYdI=،Y9S`Rn҆/hN;Ղ+MoYTU`P#Rr!Y&73cN>Iz֌C\6Y]/lk{,PSӼod 'O..ߌoU}9PDiQ /o ^ {0?'n82ora)Ջӫ%lV[lZ؁|(f Dݗ}7a{;+r0D+Jpw$śqoN[Rq1?y[֥#]!/}?oұ~hv8~\|+ՌJK!Z sWr%FX3lDT9?mqK\L3qʇkZ/`/: G I9J`PeovVh !J ,ԢTq[.];>A5e _gQ*{bl:-s+eju1OMl!^~4 |v90`-Jyque.#nu[e?}GAy`ExY'0:yk@'W \!j~*`% Ǯrk3f,k6!\Ek$ɏՕA}GvJ?fhغկnMh3W T } *7^m lqd):ygqכLnCb/d(*rb[{y-x+A,䬫AhdEo iO5S}\dWxnr27O+(Bsh>, Q=xʂj˄N ' To~!8#}JCQ%o~f÷[?Ǡ((({izP~P-pvr:Z'mu$P<~.D[ 6rYS-e-OD]?>/8/h ]σRT -T9xwŊMGJI_8\e|kί~:>uyqTI5zO5`Vxʄq% pBS$4ɣߗ]& y͑U )=+E#!n}_h;ղtS{;\Xm_U ;l*cSo샇b@xt%^$wTuLI}&^>^L҆e١8hR.iWoFbBrE6<ҧޤ?A7e{)SjÎ!iP^~lꝟ.h5IS&$jjjjʪ|'1Qy91!(0r"Ҝ:wycBJwX- ߝe Cl4#(,7~`j|omքz gVdFzoDD+E3iRvc!|7JEHd%Jv_RW֑!G,er~Jub10<[Ϧ$ W[~Ev !x>4?u>3ľӗygdH2 ~hsɵPJb3@ aFnM9IK-:yE J?B;@ZWFnP)XioV~8g_lH0Ľ-"eR#1pFZg^{B!d,PK I KIVKBh^Ow`O |~J 1ظy]~E`_¿?s",t1_2 wW 3! %!l. GhQ_nL0W@xC謨BR>z \q!)CR?-cxV`l@ Xk͕Хx1n`[5 G&xBj g5A侏>"`xCopI0A 'y"cl1o}ٌ͛=+EmtGwSᘈta1? yIpvژ]"P+ؘ7|J2_o{rscT:fp%!h-OPWP5Z(ZD{;xKj7 ¯eǨ/j6VBK2RuRQNFC-U+ *1ӥv M"`GZVpt|Px#٣9f>Œ`d(CMۻk E{NcٟjܳsT HSg+ú5XSTWbWG LH-bZkwPA#D{ݍaݬ'+iւch}ʾpn:̭!yi~-pqrH6+kARq,/хPu#QVr$Y"͜4٩h=;:yV*@=C=kt&ý>{eV5W%p5\b#8pLX䟀BKSY+hee~giͰSVo ׮Q+*U)oY7EA-SwK>V)V^S1'qd;B1|uib &5/u?"Mrx0h=pY‹Z|4^gͬE#O҄KR0TV/7W 3Bv )YI1\ }zйJ63G=F` QY5'nmx%ѱw$΂XG.of%bT+܊"Mx׸@.BDiB!:C7hӋki4 Vo~;tN] _Is{O!N l/5k!/O̭/52ZjT޼Im:^]C.{JLJoɡLp^׈v7O s]=|P||8ԏ[qc?xT {Àއs#RYtJϋF$Dzunz/=u6}roB.Dlj4r+ɑw[.1^tX-/Fz),M4˦#cZobc:Hnk"BMb-dwB.DT8b{= u8ݾ W£e ª\6vR;["ܿ~yx S9'/x[y{mGeZn?x`ͧn+g `q'DvӶHРxN0:tȞR[PCp"/Y,Cj'>(k˙]YK d K\r1!FJz^5be.Cdr&cS(WcvE|ĘN7Rzьשy~m pͲ)GGލۈ0[`$2Kb[ȅhϦ8Ҧx' *CgR/ec0m7N҄v!vk]]6E6$Ѷf _Hjĵ!Y7'omQ]=FA{\卝ڷ]N*saAzRIm^oFuXfY ޙ}var/home/core/zuul-output/logs/kubelet.log0000644000000000000000003734776115146447541017732 0ustar rootrootFeb 22 00:06:31 crc systemd[1]: Starting Kubernetes Kubelet... Feb 22 00:06:31 crc restorecon[4679]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:31 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 22 00:06:32 crc restorecon[4679]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 22 00:06:32 crc restorecon[4679]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 22 00:06:32 crc kubenswrapper[4857]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 22 00:06:32 crc kubenswrapper[4857]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 22 00:06:32 crc kubenswrapper[4857]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 22 00:06:32 crc kubenswrapper[4857]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 22 00:06:32 crc kubenswrapper[4857]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 22 00:06:32 crc kubenswrapper[4857]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.771610 4857 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779079 4857 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779109 4857 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779118 4857 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779128 4857 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779140 4857 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779151 4857 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779160 4857 feature_gate.go:330] unrecognized feature gate: Example Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779168 4857 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779178 4857 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779186 4857 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779194 4857 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779210 4857 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779218 4857 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779226 4857 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779234 4857 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779242 4857 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779250 4857 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779258 4857 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779266 4857 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779274 4857 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779282 4857 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779289 4857 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779297 4857 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779305 4857 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779313 4857 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779320 4857 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779328 4857 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779335 4857 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779343 4857 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779351 4857 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779359 4857 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779367 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779375 4857 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779383 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779391 4857 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779399 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779407 4857 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779415 4857 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779423 4857 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779431 4857 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779439 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779447 4857 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779456 4857 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779464 4857 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779471 4857 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779479 4857 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779491 4857 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779501 4857 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779510 4857 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779521 4857 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779531 4857 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779540 4857 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779548 4857 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779555 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779563 4857 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779571 4857 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779579 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779590 4857 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779598 4857 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779605 4857 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779616 4857 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779624 4857 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779632 4857 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779639 4857 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779647 4857 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779654 4857 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779662 4857 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779670 4857 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779678 4857 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779686 4857 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.779696 4857 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780745 4857 flags.go:64] FLAG: --address="0.0.0.0" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780768 4857 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780785 4857 flags.go:64] FLAG: --anonymous-auth="true" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780797 4857 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780808 4857 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780819 4857 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780830 4857 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780842 4857 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780852 4857 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780861 4857 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780873 4857 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780882 4857 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780892 4857 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780901 4857 flags.go:64] FLAG: --cgroup-root="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780934 4857 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780945 4857 flags.go:64] FLAG: --client-ca-file="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780954 4857 flags.go:64] FLAG: --cloud-config="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780964 4857 flags.go:64] FLAG: --cloud-provider="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780975 4857 flags.go:64] FLAG: --cluster-dns="[]" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780986 4857 flags.go:64] FLAG: --cluster-domain="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.780995 4857 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781005 4857 flags.go:64] FLAG: --config-dir="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781014 4857 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781023 4857 flags.go:64] FLAG: --container-log-max-files="5" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781058 4857 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781068 4857 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781077 4857 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781086 4857 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781095 4857 flags.go:64] FLAG: --contention-profiling="false" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781104 4857 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781113 4857 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781123 4857 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781133 4857 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781144 4857 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781154 4857 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781162 4857 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781171 4857 flags.go:64] FLAG: --enable-load-reader="false" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781180 4857 flags.go:64] FLAG: --enable-server="true" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781189 4857 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781201 4857 flags.go:64] FLAG: --event-burst="100" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781210 4857 flags.go:64] FLAG: --event-qps="50" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781219 4857 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781228 4857 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781237 4857 flags.go:64] FLAG: --eviction-hard="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781248 4857 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781257 4857 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781266 4857 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781275 4857 flags.go:64] FLAG: --eviction-soft="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781284 4857 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781293 4857 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781303 4857 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781311 4857 flags.go:64] FLAG: --experimental-mounter-path="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781321 4857 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781329 4857 flags.go:64] FLAG: --fail-swap-on="true" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781338 4857 flags.go:64] FLAG: --feature-gates="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781348 4857 flags.go:64] FLAG: --file-check-frequency="20s" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781358 4857 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781367 4857 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781376 4857 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781385 4857 flags.go:64] FLAG: --healthz-port="10248" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781394 4857 flags.go:64] FLAG: --help="false" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781403 4857 flags.go:64] FLAG: --hostname-override="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781412 4857 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781422 4857 flags.go:64] FLAG: --http-check-frequency="20s" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781431 4857 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781440 4857 flags.go:64] FLAG: --image-credential-provider-config="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781449 4857 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781458 4857 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781468 4857 flags.go:64] FLAG: --image-service-endpoint="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781477 4857 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781486 4857 flags.go:64] FLAG: --kube-api-burst="100" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781495 4857 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781505 4857 flags.go:64] FLAG: --kube-api-qps="50" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781514 4857 flags.go:64] FLAG: --kube-reserved="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781523 4857 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781532 4857 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781541 4857 flags.go:64] FLAG: --kubelet-cgroups="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781550 4857 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781559 4857 flags.go:64] FLAG: --lock-file="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781568 4857 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781577 4857 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781586 4857 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781604 4857 flags.go:64] FLAG: --log-json-split-stream="false" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781614 4857 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781622 4857 flags.go:64] FLAG: --log-text-split-stream="false" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781631 4857 flags.go:64] FLAG: --logging-format="text" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781640 4857 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781650 4857 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781659 4857 flags.go:64] FLAG: --manifest-url="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781668 4857 flags.go:64] FLAG: --manifest-url-header="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781679 4857 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781688 4857 flags.go:64] FLAG: --max-open-files="1000000" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781699 4857 flags.go:64] FLAG: --max-pods="110" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781709 4857 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781718 4857 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781727 4857 flags.go:64] FLAG: --memory-manager-policy="None" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781736 4857 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781745 4857 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781753 4857 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781763 4857 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781782 4857 flags.go:64] FLAG: --node-status-max-images="50" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781791 4857 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781800 4857 flags.go:64] FLAG: --oom-score-adj="-999" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781809 4857 flags.go:64] FLAG: --pod-cidr="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781820 4857 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781834 4857 flags.go:64] FLAG: --pod-manifest-path="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781843 4857 flags.go:64] FLAG: --pod-max-pids="-1" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781852 4857 flags.go:64] FLAG: --pods-per-core="0" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781861 4857 flags.go:64] FLAG: --port="10250" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781871 4857 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781879 4857 flags.go:64] FLAG: --provider-id="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781888 4857 flags.go:64] FLAG: --qos-reserved="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781897 4857 flags.go:64] FLAG: --read-only-port="10255" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781906 4857 flags.go:64] FLAG: --register-node="true" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781916 4857 flags.go:64] FLAG: --register-schedulable="true" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781924 4857 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781939 4857 flags.go:64] FLAG: --registry-burst="10" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781948 4857 flags.go:64] FLAG: --registry-qps="5" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781957 4857 flags.go:64] FLAG: --reserved-cpus="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781965 4857 flags.go:64] FLAG: --reserved-memory="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781977 4857 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781986 4857 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.781995 4857 flags.go:64] FLAG: --rotate-certificates="false" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782004 4857 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782013 4857 flags.go:64] FLAG: --runonce="false" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782022 4857 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782031 4857 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782068 4857 flags.go:64] FLAG: --seccomp-default="false" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782077 4857 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782087 4857 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782096 4857 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782106 4857 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782115 4857 flags.go:64] FLAG: --storage-driver-password="root" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782124 4857 flags.go:64] FLAG: --storage-driver-secure="false" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782132 4857 flags.go:64] FLAG: --storage-driver-table="stats" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782141 4857 flags.go:64] FLAG: --storage-driver-user="root" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782151 4857 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782161 4857 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782170 4857 flags.go:64] FLAG: --system-cgroups="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782180 4857 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782195 4857 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782205 4857 flags.go:64] FLAG: --tls-cert-file="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782214 4857 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782224 4857 flags.go:64] FLAG: --tls-min-version="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782233 4857 flags.go:64] FLAG: --tls-private-key-file="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782242 4857 flags.go:64] FLAG: --topology-manager-policy="none" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782252 4857 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782261 4857 flags.go:64] FLAG: --topology-manager-scope="container" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782270 4857 flags.go:64] FLAG: --v="2" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782282 4857 flags.go:64] FLAG: --version="false" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782294 4857 flags.go:64] FLAG: --vmodule="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782305 4857 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.782314 4857 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782522 4857 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782534 4857 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782546 4857 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782556 4857 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782566 4857 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782574 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782583 4857 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782592 4857 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782601 4857 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782610 4857 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782617 4857 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782626 4857 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782634 4857 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782642 4857 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782650 4857 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782658 4857 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782666 4857 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782673 4857 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782681 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782690 4857 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782697 4857 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782705 4857 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782713 4857 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782722 4857 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782730 4857 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782741 4857 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782750 4857 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782759 4857 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782767 4857 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782775 4857 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782783 4857 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782791 4857 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782801 4857 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782810 4857 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782819 4857 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782828 4857 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782837 4857 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782845 4857 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782853 4857 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782860 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782869 4857 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782876 4857 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782884 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782892 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782900 4857 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782908 4857 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782916 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782924 4857 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782931 4857 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782939 4857 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782947 4857 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782955 4857 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782963 4857 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782971 4857 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782978 4857 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782987 4857 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.782995 4857 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.783004 4857 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.783012 4857 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.783020 4857 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.783028 4857 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.783062 4857 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.783070 4857 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.783078 4857 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.783086 4857 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.783094 4857 feature_gate.go:330] unrecognized feature gate: Example Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.783102 4857 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.783110 4857 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.783118 4857 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.783128 4857 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.783137 4857 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.783161 4857 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.794060 4857 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.794104 4857 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794213 4857 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794222 4857 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794226 4857 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794231 4857 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794235 4857 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794239 4857 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794243 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794247 4857 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794250 4857 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794256 4857 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794264 4857 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794269 4857 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794273 4857 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794278 4857 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794283 4857 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794289 4857 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794293 4857 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794298 4857 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794303 4857 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794307 4857 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794311 4857 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794315 4857 feature_gate.go:330] unrecognized feature gate: Example Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794318 4857 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794322 4857 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794325 4857 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794329 4857 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794335 4857 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794339 4857 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794342 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794346 4857 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794350 4857 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794354 4857 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794358 4857 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794362 4857 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794372 4857 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794376 4857 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794379 4857 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794383 4857 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794386 4857 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794391 4857 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794395 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794399 4857 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794402 4857 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794406 4857 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794409 4857 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794412 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794416 4857 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794427 4857 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794431 4857 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794434 4857 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794439 4857 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794443 4857 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794447 4857 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794451 4857 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794454 4857 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794458 4857 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794461 4857 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794465 4857 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794471 4857 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794476 4857 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794480 4857 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794484 4857 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794488 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794495 4857 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794500 4857 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794504 4857 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794508 4857 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794513 4857 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794517 4857 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794520 4857 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794528 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.794537 4857 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794691 4857 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794701 4857 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794707 4857 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794711 4857 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794716 4857 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794720 4857 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794724 4857 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794729 4857 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794733 4857 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794737 4857 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794741 4857 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794746 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794751 4857 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794755 4857 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794759 4857 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794763 4857 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794768 4857 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794774 4857 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794779 4857 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794784 4857 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794790 4857 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794795 4857 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794799 4857 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794803 4857 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794809 4857 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794814 4857 feature_gate.go:330] unrecognized feature gate: Example Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794819 4857 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794825 4857 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794830 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794835 4857 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794840 4857 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794846 4857 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794851 4857 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794856 4857 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794861 4857 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794867 4857 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794872 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794877 4857 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794882 4857 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794887 4857 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794892 4857 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794897 4857 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794901 4857 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794906 4857 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794910 4857 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794914 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794918 4857 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794922 4857 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794926 4857 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794930 4857 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794934 4857 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794937 4857 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794942 4857 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794946 4857 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794950 4857 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794954 4857 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794958 4857 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794962 4857 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794966 4857 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794970 4857 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794974 4857 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794977 4857 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794981 4857 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794985 4857 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794989 4857 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794993 4857 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.794997 4857 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.795001 4857 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.795004 4857 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.795009 4857 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.795013 4857 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.795021 4857 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.795255 4857 server.go:940] "Client rotation is on, will bootstrap in background" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.800098 4857 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.800191 4857 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.801545 4857 server.go:997] "Starting client certificate rotation" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.801569 4857 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.802551 4857 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-08 02:00:22.742261242 +0000 UTC Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.802646 4857 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.838729 4857 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 22 00:06:32 crc kubenswrapper[4857]: E0222 00:06:32.841548 4857 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.843479 4857 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.872746 4857 log.go:25] "Validated CRI v1 runtime API" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.915249 4857 log.go:25] "Validated CRI v1 image API" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.918140 4857 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.925007 4857 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-22-00-02-08-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.925141 4857 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.953471 4857 manager.go:217] Machine: {Timestamp:2026-02-22 00:06:32.949765655 +0000 UTC m=+0.588495008 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:f5971b29-a603-4824-855e-72c4a0612518 BootID:2a183360-3a9c-475d-a3f6-4b5064ea42f5 Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:f7:12:0c Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:f7:12:0c Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:c5:19:34 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:22:c3:2c Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:13:06:d1 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:96:81:c9 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:ca:d3:a8:cf:14:d2 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:a6:b2:e5:91:8c:c7 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.953863 4857 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.954240 4857 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.956975 4857 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.957351 4857 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.957409 4857 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.958441 4857 topology_manager.go:138] "Creating topology manager with none policy" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.958487 4857 container_manager_linux.go:303] "Creating device plugin manager" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.959443 4857 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.959478 4857 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.960594 4857 state_mem.go:36] "Initialized new in-memory state store" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.960747 4857 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.965112 4857 kubelet.go:418] "Attempting to sync node with API server" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.965150 4857 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.965183 4857 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.965206 4857 kubelet.go:324] "Adding apiserver pod source" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.965228 4857 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.972216 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:06:32 crc kubenswrapper[4857]: E0222 00:06:32.972363 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.972441 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:06:32 crc kubenswrapper[4857]: E0222 00:06:32.972672 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.974299 4857 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.975720 4857 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.977651 4857 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.979567 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.979597 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.979608 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.979634 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.979648 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.979655 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.979664 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.979676 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.979686 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.979694 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.980363 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.980438 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.981516 4857 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.982188 4857 server.go:1280] "Started kubelet" Feb 22 00:06:32 crc systemd[1]: Started Kubernetes Kubelet. Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.984326 4857 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.984370 4857 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.984326 4857 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.984842 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.986057 4857 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.986301 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 20:46:58.282757881 +0000 UTC Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.986406 4857 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.986427 4857 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 22 00:06:32 crc kubenswrapper[4857]: E0222 00:06:32.986429 4857 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.986589 4857 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.986653 4857 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 22 00:06:32 crc kubenswrapper[4857]: E0222 00:06:32.987489 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="200ms" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.988363 4857 factory.go:153] Registering CRI-O factory Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.988394 4857 factory.go:221] Registration of the crio container factory successfully Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.988474 4857 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.988486 4857 factory.go:55] Registering systemd factory Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.988494 4857 factory.go:221] Registration of the systemd container factory successfully Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.988515 4857 factory.go:103] Registering Raw factory Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.988532 4857 manager.go:1196] Started watching for new ooms in manager Feb 22 00:06:32 crc kubenswrapper[4857]: W0222 00:06:32.989840 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:06:32 crc kubenswrapper[4857]: E0222 00:06:32.989978 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.990559 4857 server.go:460] "Adding debug handlers to kubelet server" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.993886 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.993962 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.993979 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.993994 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994008 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994021 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994061 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994074 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994089 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994101 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994113 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994126 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994138 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994152 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994163 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994175 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994191 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994205 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994217 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994229 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994244 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994258 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994270 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994289 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994305 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994321 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994338 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994354 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994369 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994384 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994398 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994412 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994432 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994447 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994461 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994474 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994487 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994502 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994516 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994531 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994546 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994560 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994573 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994587 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994620 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994640 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994653 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994666 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994684 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994705 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994719 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994733 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994753 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994768 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.994783 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.996316 4857 manager.go:319] Starting recovery of all containers Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.998794 4857 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.998841 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.998862 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.998880 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.998896 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.998914 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.998935 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.998950 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.998966 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.998980 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.998994 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.999008 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.999023 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.999052 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.999069 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.999084 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.999099 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.999113 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.999131 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.999145 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 22 00:06:32 crc kubenswrapper[4857]: I0222 00:06:32.999162 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:32.999179 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:32.999195 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:32.999211 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:32.999225 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.001005 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.001031 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.001085 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.001102 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.001124 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.001140 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.001156 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.002955 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.003201 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.003294 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: E0222 00:06:33.003448 4857 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.180:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189668daf195cc6d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-22 00:06:32.982154349 +0000 UTC m=+0.620883602,LastTimestamp:2026-02-22 00:06:32.982154349 +0000 UTC m=+0.620883602,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.003694 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.003803 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.003939 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.004225 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.004313 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.004404 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.004496 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.004646 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.004695 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.004783 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.004883 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.004965 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.005076 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.005164 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.005251 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.005370 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.005484 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.005581 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.005737 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.005851 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.005940 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.006008 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.006119 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.006162 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.006215 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.006249 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.006347 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.006386 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.006434 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.006466 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.006510 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.006555 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.006587 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.006637 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.006732 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.006778 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.006865 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.006930 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.006977 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.007084 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.007126 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.007216 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.007286 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.007370 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.007401 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.007466 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.007553 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.007590 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.007671 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008190 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008244 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008275 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008294 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008311 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008332 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008350 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008371 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008387 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008404 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008424 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008444 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008469 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008485 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008503 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008531 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008548 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008577 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008595 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008610 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008641 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008657 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008876 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008914 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008940 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008976 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.008994 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009017 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009049 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009068 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009101 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009120 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009147 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009411 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009442 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009468 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009487 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009509 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009528 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009547 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009568 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009589 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009610 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009626 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009643 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009663 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009683 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.009704 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.010454 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.010838 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.010886 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.010906 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.010930 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.010948 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.010969 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.010994 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.011011 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.011048 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.011072 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.011104 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.012112 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.012187 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.012221 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.012257 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.012332 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.013358 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.013385 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.013402 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.013417 4857 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.013431 4857 reconstruct.go:97] "Volume reconstruction finished" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.013445 4857 reconciler.go:26] "Reconciler: start to sync state" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.023964 4857 manager.go:324] Recovery completed Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.035640 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.038301 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.038343 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.038354 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.040773 4857 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.040813 4857 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.040848 4857 state_mem.go:36] "Initialized new in-memory state store" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.058112 4857 policy_none.go:49] "None policy: Start" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.062657 4857 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.062906 4857 state_mem.go:35] "Initializing new in-memory state store" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.073880 4857 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.076217 4857 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.076260 4857 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.076299 4857 kubelet.go:2335] "Starting kubelet main sync loop" Feb 22 00:06:33 crc kubenswrapper[4857]: E0222 00:06:33.076351 4857 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 22 00:06:33 crc kubenswrapper[4857]: W0222 00:06:33.077643 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:06:33 crc kubenswrapper[4857]: E0222 00:06:33.077711 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Feb 22 00:06:33 crc kubenswrapper[4857]: E0222 00:06:33.087137 4857 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.125790 4857 manager.go:334] "Starting Device Plugin manager" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.125876 4857 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.125905 4857 server.go:79] "Starting device plugin registration server" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.126996 4857 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.127061 4857 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.127371 4857 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.127505 4857 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.127527 4857 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 22 00:06:33 crc kubenswrapper[4857]: E0222 00:06:33.141235 4857 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.177489 4857 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.177598 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.179093 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.179206 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.179236 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.179679 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.179793 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.179832 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.180886 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.180918 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.180928 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.182138 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.182209 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.182236 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.182584 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.182900 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.183004 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.184516 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.184543 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.184554 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.184695 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.184851 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.184915 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.185456 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.185485 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.185497 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.185632 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.185834 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.185883 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.187501 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.187549 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.187561 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.187905 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.187927 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.187939 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.188127 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.188154 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.188167 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.188227 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.188312 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.188334 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:33 crc kubenswrapper[4857]: E0222 00:06:33.189142 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="400ms" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.191998 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.193188 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.194971 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.195006 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.195016 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.218594 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.218632 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.218654 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.218670 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.218688 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.218713 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.218730 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.218747 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.218764 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.218798 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.218816 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.218836 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.218855 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.218875 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.218957 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.227213 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.228650 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.228860 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.228996 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.229317 4857 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 22 00:06:33 crc kubenswrapper[4857]: E0222 00:06:33.230020 4857 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.319984 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320091 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320119 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320145 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320168 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320190 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320210 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320230 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320253 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320319 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320346 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320378 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320383 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320454 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320515 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320555 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320571 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320603 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320607 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320627 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320281 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320678 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320706 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320404 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320759 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320770 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320801 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320865 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320907 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.320954 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.431017 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.432749 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.432801 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.432812 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.432849 4857 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 22 00:06:33 crc kubenswrapper[4857]: E0222 00:06:33.433419 4857 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.516318 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.525169 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.545024 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.574453 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: W0222 00:06:33.576227 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-2d88c217b98891c1b2e63198a9d77ee5d66df527ff49a4f3aea03541e5f00ffe WatchSource:0}: Error finding container 2d88c217b98891c1b2e63198a9d77ee5d66df527ff49a4f3aea03541e5f00ffe: Status 404 returned error can't find the container with id 2d88c217b98891c1b2e63198a9d77ee5d66df527ff49a4f3aea03541e5f00ffe Feb 22 00:06:33 crc kubenswrapper[4857]: W0222 00:06:33.581327 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-fbaa751e016ac0b5097598918c462025e3bf6b181315a77ee2d97263cb76db5d WatchSource:0}: Error finding container fbaa751e016ac0b5097598918c462025e3bf6b181315a77ee2d97263cb76db5d: Status 404 returned error can't find the container with id fbaa751e016ac0b5097598918c462025e3bf6b181315a77ee2d97263cb76db5d Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.581607 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 22 00:06:33 crc kubenswrapper[4857]: E0222 00:06:33.590106 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="800ms" Feb 22 00:06:33 crc kubenswrapper[4857]: W0222 00:06:33.593430 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-b5fc5070ebc7618d8dd2ec9becddc31168c4c3cc0aa70538ea4a1a117cbc6b28 WatchSource:0}: Error finding container b5fc5070ebc7618d8dd2ec9becddc31168c4c3cc0aa70538ea4a1a117cbc6b28: Status 404 returned error can't find the container with id b5fc5070ebc7618d8dd2ec9becddc31168c4c3cc0aa70538ea4a1a117cbc6b28 Feb 22 00:06:33 crc kubenswrapper[4857]: W0222 00:06:33.608389 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-2c9089aac5192613ed9dac88f46aa39d4b29ce5e1731b1b9382f1eb15edc411c WatchSource:0}: Error finding container 2c9089aac5192613ed9dac88f46aa39d4b29ce5e1731b1b9382f1eb15edc411c: Status 404 returned error can't find the container with id 2c9089aac5192613ed9dac88f46aa39d4b29ce5e1731b1b9382f1eb15edc411c Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.833616 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.836485 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.836556 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.836583 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.836634 4857 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 22 00:06:33 crc kubenswrapper[4857]: E0222 00:06:33.837348 4857 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Feb 22 00:06:33 crc kubenswrapper[4857]: W0222 00:06:33.895408 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:06:33 crc kubenswrapper[4857]: E0222 00:06:33.895579 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.986711 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 13:59:08.750819958 +0000 UTC Feb 22 00:06:33 crc kubenswrapper[4857]: I0222 00:06:33.987528 4857 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:06:34 crc kubenswrapper[4857]: I0222 00:06:34.082261 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"fbaa751e016ac0b5097598918c462025e3bf6b181315a77ee2d97263cb76db5d"} Feb 22 00:06:34 crc kubenswrapper[4857]: I0222 00:06:34.085357 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2c9089aac5192613ed9dac88f46aa39d4b29ce5e1731b1b9382f1eb15edc411c"} Feb 22 00:06:34 crc kubenswrapper[4857]: I0222 00:06:34.087982 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2f9ffaa66b9271f91888eb8734012b65752b023e998479e67e28323aa7ddc439"} Feb 22 00:06:34 crc kubenswrapper[4857]: I0222 00:06:34.089415 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b5fc5070ebc7618d8dd2ec9becddc31168c4c3cc0aa70538ea4a1a117cbc6b28"} Feb 22 00:06:34 crc kubenswrapper[4857]: I0222 00:06:34.090866 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2d88c217b98891c1b2e63198a9d77ee5d66df527ff49a4f3aea03541e5f00ffe"} Feb 22 00:06:34 crc kubenswrapper[4857]: W0222 00:06:34.121087 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:06:34 crc kubenswrapper[4857]: E0222 00:06:34.121240 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Feb 22 00:06:34 crc kubenswrapper[4857]: E0222 00:06:34.391015 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="1.6s" Feb 22 00:06:34 crc kubenswrapper[4857]: W0222 00:06:34.537149 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:06:34 crc kubenswrapper[4857]: E0222 00:06:34.537307 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Feb 22 00:06:34 crc kubenswrapper[4857]: W0222 00:06:34.543319 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:06:34 crc kubenswrapper[4857]: E0222 00:06:34.543492 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Feb 22 00:06:34 crc kubenswrapper[4857]: I0222 00:06:34.638477 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:34 crc kubenswrapper[4857]: I0222 00:06:34.640851 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:34 crc kubenswrapper[4857]: I0222 00:06:34.640910 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:34 crc kubenswrapper[4857]: I0222 00:06:34.640920 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:34 crc kubenswrapper[4857]: I0222 00:06:34.640952 4857 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 22 00:06:34 crc kubenswrapper[4857]: E0222 00:06:34.641511 4857 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Feb 22 00:06:34 crc kubenswrapper[4857]: I0222 00:06:34.879937 4857 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 22 00:06:34 crc kubenswrapper[4857]: E0222 00:06:34.881352 4857 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Feb 22 00:06:34 crc kubenswrapper[4857]: I0222 00:06:34.987112 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 19:06:46.151726304 +0000 UTC Feb 22 00:06:34 crc kubenswrapper[4857]: I0222 00:06:34.987695 4857 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.096785 4857 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711" exitCode=0 Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.096953 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711"} Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.097013 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.099084 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.099146 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.099168 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.100756 4857 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f" exitCode=0 Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.100901 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.100899 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f"} Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.101963 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.102030 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.102121 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.103562 4857 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024" exitCode=0 Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.103918 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.104215 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024"} Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.105411 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.107428 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.107459 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.107472 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.107476 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.107522 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.107553 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.117701 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0"} Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.117788 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162"} Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.117820 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11"} Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.121841 4857 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="2c1063303cf81fdf58d084698736e9405e23a7fe947708b457c67ad8677fcf11" exitCode=0 Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.121893 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"2c1063303cf81fdf58d084698736e9405e23a7fe947708b457c67ad8677fcf11"} Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.122051 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.123246 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.123300 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.123318 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.987062 4857 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:06:35 crc kubenswrapper[4857]: I0222 00:06:35.987464 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 02:23:30.89585565 +0000 UTC Feb 22 00:06:35 crc kubenswrapper[4857]: E0222 00:06:35.991589 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="3.2s" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.127170 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa"} Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.127227 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f"} Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.127241 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45"} Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.127253 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df"} Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.128620 4857 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211" exitCode=0 Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.128676 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211"} Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.128726 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.129744 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.129775 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.129787 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.131794 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de"} Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.131869 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.132641 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.132666 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.132677 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.133774 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a3453545318207b5b5c34f8ce00f1d50b9d33674903fb173c2506fc5332a21eb"} Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.133786 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.134914 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.134942 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.134952 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.136266 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9920d33f76b36664452e744820c9fb71956c5773ca81b750d6202ed6cf08c4d6"} Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.136296 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8f62ed2ecd8930d7f3e899f7891c87b94f1e117a23588b2ce65581a3bd37f225"} Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.136310 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"21a6a9ee3b161774e7dffa093ffbff0c3fb4202d2ca7a0351ae11bd2d5de064c"} Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.136321 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.136871 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.136888 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.136896 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.242229 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.243651 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.243688 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.243701 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.243728 4857 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 22 00:06:36 crc kubenswrapper[4857]: E0222 00:06:36.244146 4857 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Feb 22 00:06:36 crc kubenswrapper[4857]: W0222 00:06:36.473661 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:06:36 crc kubenswrapper[4857]: E0222 00:06:36.473792 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Feb 22 00:06:36 crc kubenswrapper[4857]: W0222 00:06:36.562391 4857 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:06:36 crc kubenswrapper[4857]: E0222 00:06:36.562522 4857 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.594590 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.602393 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:06:36 crc kubenswrapper[4857]: I0222 00:06:36.988231 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 12:45:31.165473476 +0000 UTC Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.152857 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"43984627e32a20f5c61d14d0badf2782f948f3817107a051a8b61d0c8aaa1c2c"} Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.152926 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.154427 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.154516 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.154551 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.155942 4857 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7" exitCode=0 Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.156150 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7"} Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.156250 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.156297 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.156366 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.157411 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.157514 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.157804 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.157991 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.158073 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.158179 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.158225 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.158248 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.158178 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.158304 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.158322 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.159586 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.159639 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.159657 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:37 crc kubenswrapper[4857]: I0222 00:06:37.988822 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 01:40:19.672007003 +0000 UTC Feb 22 00:06:38 crc kubenswrapper[4857]: I0222 00:06:38.162602 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 22 00:06:38 crc kubenswrapper[4857]: I0222 00:06:38.162658 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:38 crc kubenswrapper[4857]: I0222 00:06:38.162703 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9"} Feb 22 00:06:38 crc kubenswrapper[4857]: I0222 00:06:38.162772 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c"} Feb 22 00:06:38 crc kubenswrapper[4857]: I0222 00:06:38.162795 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457"} Feb 22 00:06:38 crc kubenswrapper[4857]: I0222 00:06:38.162808 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16"} Feb 22 00:06:38 crc kubenswrapper[4857]: I0222 00:06:38.162860 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 22 00:06:38 crc kubenswrapper[4857]: I0222 00:06:38.162936 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:38 crc kubenswrapper[4857]: I0222 00:06:38.166451 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:38 crc kubenswrapper[4857]: I0222 00:06:38.166497 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:38 crc kubenswrapper[4857]: I0222 00:06:38.166516 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:38 crc kubenswrapper[4857]: I0222 00:06:38.166549 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:38 crc kubenswrapper[4857]: I0222 00:06:38.166596 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:38 crc kubenswrapper[4857]: I0222 00:06:38.166652 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:38 crc kubenswrapper[4857]: I0222 00:06:38.303205 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:06:38 crc kubenswrapper[4857]: I0222 00:06:38.984090 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:06:38 crc kubenswrapper[4857]: I0222 00:06:38.988983 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 11:34:33.449706417 +0000 UTC Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.171132 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.171089 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77"} Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.171186 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.171232 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.171297 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.171338 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.172972 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.173096 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.173120 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.173255 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.173302 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.173326 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.174202 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.174226 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.174238 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.179544 4857 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.444713 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.447295 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.447364 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.447391 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.447435 4857 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 22 00:06:39 crc kubenswrapper[4857]: I0222 00:06:39.989987 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 17:21:13.976714258 +0000 UTC Feb 22 00:06:40 crc kubenswrapper[4857]: I0222 00:06:40.081682 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:06:40 crc kubenswrapper[4857]: I0222 00:06:40.172970 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:40 crc kubenswrapper[4857]: I0222 00:06:40.173015 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:40 crc kubenswrapper[4857]: I0222 00:06:40.174127 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:40 crc kubenswrapper[4857]: I0222 00:06:40.174160 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:40 crc kubenswrapper[4857]: I0222 00:06:40.174170 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:40 crc kubenswrapper[4857]: I0222 00:06:40.174349 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:40 crc kubenswrapper[4857]: I0222 00:06:40.174368 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:40 crc kubenswrapper[4857]: I0222 00:06:40.174379 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:40 crc kubenswrapper[4857]: I0222 00:06:40.308984 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 22 00:06:40 crc kubenswrapper[4857]: I0222 00:06:40.991125 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 21:12:15.778419772 +0000 UTC Feb 22 00:06:41 crc kubenswrapper[4857]: I0222 00:06:41.157838 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:06:41 crc kubenswrapper[4857]: I0222 00:06:41.158294 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:41 crc kubenswrapper[4857]: I0222 00:06:41.161124 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:41 crc kubenswrapper[4857]: I0222 00:06:41.161204 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:41 crc kubenswrapper[4857]: I0222 00:06:41.161227 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:41 crc kubenswrapper[4857]: I0222 00:06:41.175199 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:41 crc kubenswrapper[4857]: I0222 00:06:41.176805 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:41 crc kubenswrapper[4857]: I0222 00:06:41.176855 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:41 crc kubenswrapper[4857]: I0222 00:06:41.176865 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:41 crc kubenswrapper[4857]: I0222 00:06:41.698862 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:06:41 crc kubenswrapper[4857]: I0222 00:06:41.699085 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:41 crc kubenswrapper[4857]: I0222 00:06:41.700827 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:41 crc kubenswrapper[4857]: I0222 00:06:41.700866 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:41 crc kubenswrapper[4857]: I0222 00:06:41.700882 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:41 crc kubenswrapper[4857]: I0222 00:06:41.991729 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 18:03:14.445798516 +0000 UTC Feb 22 00:06:42 crc kubenswrapper[4857]: I0222 00:06:42.436409 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 22 00:06:42 crc kubenswrapper[4857]: I0222 00:06:42.436789 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:42 crc kubenswrapper[4857]: I0222 00:06:42.438685 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:42 crc kubenswrapper[4857]: I0222 00:06:42.438737 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:42 crc kubenswrapper[4857]: I0222 00:06:42.438758 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:42 crc kubenswrapper[4857]: I0222 00:06:42.992295 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 02:15:40.197226576 +0000 UTC Feb 22 00:06:43 crc kubenswrapper[4857]: E0222 00:06:43.141622 4857 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 22 00:06:43 crc kubenswrapper[4857]: I0222 00:06:43.393689 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 22 00:06:43 crc kubenswrapper[4857]: I0222 00:06:43.393928 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:43 crc kubenswrapper[4857]: I0222 00:06:43.395317 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:43 crc kubenswrapper[4857]: I0222 00:06:43.395541 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:43 crc kubenswrapper[4857]: I0222 00:06:43.395550 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:43 crc kubenswrapper[4857]: I0222 00:06:43.766363 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:06:43 crc kubenswrapper[4857]: I0222 00:06:43.766525 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:43 crc kubenswrapper[4857]: I0222 00:06:43.767966 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:43 crc kubenswrapper[4857]: I0222 00:06:43.768113 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:43 crc kubenswrapper[4857]: I0222 00:06:43.768139 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:43 crc kubenswrapper[4857]: I0222 00:06:43.992846 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 15:25:49.204263592 +0000 UTC Feb 22 00:06:44 crc kubenswrapper[4857]: I0222 00:06:44.992988 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 03:12:23.545001363 +0000 UTC Feb 22 00:06:45 crc kubenswrapper[4857]: I0222 00:06:45.993218 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 22:35:46.898076764 +0000 UTC Feb 22 00:06:46 crc kubenswrapper[4857]: I0222 00:06:46.767357 4857 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 22 00:06:46 crc kubenswrapper[4857]: I0222 00:06:46.767430 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 22 00:06:46 crc kubenswrapper[4857]: I0222 00:06:46.987629 4857 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 22 00:06:46 crc kubenswrapper[4857]: I0222 00:06:46.993695 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 15:54:23.133742071 +0000 UTC Feb 22 00:06:47 crc kubenswrapper[4857]: I0222 00:06:47.194835 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 22 00:06:47 crc kubenswrapper[4857]: I0222 00:06:47.196824 4857 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="43984627e32a20f5c61d14d0badf2782f948f3817107a051a8b61d0c8aaa1c2c" exitCode=255 Feb 22 00:06:47 crc kubenswrapper[4857]: I0222 00:06:47.196877 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"43984627e32a20f5c61d14d0badf2782f948f3817107a051a8b61d0c8aaa1c2c"} Feb 22 00:06:47 crc kubenswrapper[4857]: I0222 00:06:47.197099 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:47 crc kubenswrapper[4857]: I0222 00:06:47.197988 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:47 crc kubenswrapper[4857]: I0222 00:06:47.198029 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:47 crc kubenswrapper[4857]: I0222 00:06:47.198073 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:47 crc kubenswrapper[4857]: I0222 00:06:47.198690 4857 scope.go:117] "RemoveContainer" containerID="43984627e32a20f5c61d14d0badf2782f948f3817107a051a8b61d0c8aaa1c2c" Feb 22 00:06:47 crc kubenswrapper[4857]: I0222 00:06:47.230167 4857 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 22 00:06:47 crc kubenswrapper[4857]: I0222 00:06:47.230236 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 22 00:06:47 crc kubenswrapper[4857]: I0222 00:06:47.241818 4857 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Feb 22 00:06:47 crc kubenswrapper[4857]: I0222 00:06:47.241893 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 22 00:06:47 crc kubenswrapper[4857]: I0222 00:06:47.994634 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 21:16:53.079201953 +0000 UTC Feb 22 00:06:48 crc kubenswrapper[4857]: I0222 00:06:48.202784 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 22 00:06:48 crc kubenswrapper[4857]: I0222 00:06:48.205887 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253"} Feb 22 00:06:48 crc kubenswrapper[4857]: I0222 00:06:48.206149 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:48 crc kubenswrapper[4857]: I0222 00:06:48.222971 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:48 crc kubenswrapper[4857]: I0222 00:06:48.223031 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:48 crc kubenswrapper[4857]: I0222 00:06:48.223076 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:48 crc kubenswrapper[4857]: I0222 00:06:48.990415 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:06:48 crc kubenswrapper[4857]: I0222 00:06:48.995219 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 15:08:21.194833267 +0000 UTC Feb 22 00:06:49 crc kubenswrapper[4857]: I0222 00:06:49.208741 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:49 crc kubenswrapper[4857]: I0222 00:06:49.208927 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:06:49 crc kubenswrapper[4857]: I0222 00:06:49.209776 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:49 crc kubenswrapper[4857]: I0222 00:06:49.209820 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:49 crc kubenswrapper[4857]: I0222 00:06:49.209838 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:49 crc kubenswrapper[4857]: I0222 00:06:49.213530 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:06:49 crc kubenswrapper[4857]: I0222 00:06:49.996097 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 02:56:03.38004407 +0000 UTC Feb 22 00:06:50 crc kubenswrapper[4857]: I0222 00:06:50.091798 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:06:50 crc kubenswrapper[4857]: I0222 00:06:50.092014 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:50 crc kubenswrapper[4857]: I0222 00:06:50.093821 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:50 crc kubenswrapper[4857]: I0222 00:06:50.093863 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:50 crc kubenswrapper[4857]: I0222 00:06:50.093882 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:50 crc kubenswrapper[4857]: I0222 00:06:50.211388 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:50 crc kubenswrapper[4857]: I0222 00:06:50.213000 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:50 crc kubenswrapper[4857]: I0222 00:06:50.213087 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:50 crc kubenswrapper[4857]: I0222 00:06:50.213111 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:50 crc kubenswrapper[4857]: I0222 00:06:50.997130 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 10:30:32.996461776 +0000 UTC Feb 22 00:06:51 crc kubenswrapper[4857]: I0222 00:06:51.214888 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:51 crc kubenswrapper[4857]: I0222 00:06:51.216349 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:51 crc kubenswrapper[4857]: I0222 00:06:51.216393 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:51 crc kubenswrapper[4857]: I0222 00:06:51.216415 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:51 crc kubenswrapper[4857]: I0222 00:06:51.998077 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 04:11:40.034736256 +0000 UTC Feb 22 00:06:52 crc kubenswrapper[4857]: E0222 00:06:52.227188 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.231163 4857 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.231503 4857 trace.go:236] Trace[552421439]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Feb-2026 00:06:37.609) (total time: 14621ms): Feb 22 00:06:52 crc kubenswrapper[4857]: Trace[552421439]: ---"Objects listed" error: 14621ms (00:06:52.231) Feb 22 00:06:52 crc kubenswrapper[4857]: Trace[552421439]: [14.621642216s] [14.621642216s] END Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.231968 4857 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.232775 4857 trace.go:236] Trace[63400048]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Feb-2026 00:06:37.310) (total time: 14922ms): Feb 22 00:06:52 crc kubenswrapper[4857]: Trace[63400048]: ---"Objects listed" error: 14922ms (00:06:52.232) Feb 22 00:06:52 crc kubenswrapper[4857]: Trace[63400048]: [14.922439264s] [14.922439264s] END Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.232912 4857 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 22 00:06:52 crc kubenswrapper[4857]: E0222 00:06:52.233320 4857 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.234771 4857 trace.go:236] Trace[1865839772]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (22-Feb-2026 00:06:41.932) (total time: 10302ms): Feb 22 00:06:52 crc kubenswrapper[4857]: Trace[1865839772]: ---"Objects listed" error: 10302ms (00:06:52.234) Feb 22 00:06:52 crc kubenswrapper[4857]: Trace[1865839772]: [10.302216765s] [10.302216765s] END Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.234886 4857 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.234813 4857 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.254774 4857 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.977649 4857 apiserver.go:52] "Watching apiserver" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.983205 4857 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.983693 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.984530 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.984911 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.984950 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:06:52 crc kubenswrapper[4857]: E0222 00:06:52.985092 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:06:52 crc kubenswrapper[4857]: E0222 00:06:52.985132 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.985420 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.985515 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.985541 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 22 00:06:52 crc kubenswrapper[4857]: E0222 00:06:52.985758 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.987456 4857 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.989125 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.990417 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.990783 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.990944 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.990955 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.991070 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.991510 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.992103 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 22 00:06:52 crc kubenswrapper[4857]: I0222 00:06:52.993703 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.000236 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 21:28:55.451549109 +0000 UTC Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.029924 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.041713 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.041768 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.041796 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.041828 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.041856 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.041877 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.041899 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.041927 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042085 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042150 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042173 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042195 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042218 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042239 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042263 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042284 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042332 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042356 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042382 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042407 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042431 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042487 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042508 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042562 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042585 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042607 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042631 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042655 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042676 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042698 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042721 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042741 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042763 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042784 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042808 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042832 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042854 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042877 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042899 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042919 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042964 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.042986 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043007 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043029 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043070 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043090 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043110 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043131 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043152 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043173 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043194 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043215 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043236 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043257 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043277 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043300 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043321 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043342 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043363 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043385 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043407 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043428 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043479 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043503 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043504 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043526 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043631 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043675 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043725 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043766 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043807 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043818 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043843 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043882 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043917 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043952 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.043986 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044020 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044138 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044194 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044271 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044331 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044380 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044420 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044456 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044492 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044527 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044566 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044603 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044637 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044671 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044743 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044830 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044872 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044902 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044927 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044951 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044972 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.044995 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045018 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045262 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045298 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045306 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045339 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045363 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045387 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045409 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045430 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045451 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045472 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045492 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045513 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045534 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045557 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045560 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045581 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045605 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045628 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045649 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045672 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045694 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045718 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045741 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045763 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045784 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045806 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045829 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045851 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045873 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045896 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045920 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045941 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045962 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045983 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046010 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046030 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046085 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046114 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046144 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046167 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046191 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046213 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046233 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046255 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046277 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046302 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046329 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046360 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046396 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046428 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046479 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046512 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046547 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046583 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046620 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046651 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046685 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046716 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046753 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046787 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046820 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046855 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046890 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046922 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046957 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046992 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047028 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047085 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047115 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047147 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047181 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047214 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047249 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047283 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047320 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047351 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047390 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047423 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047455 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047486 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047515 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047547 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047578 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047609 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047641 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047670 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047715 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047746 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047782 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047818 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047847 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047874 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047902 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047933 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047971 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048003 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048056 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048092 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048126 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048160 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048223 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048274 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048345 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048386 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048422 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048459 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048497 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048524 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048548 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048572 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048598 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048624 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048663 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.053614 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.055438 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.057753 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045670 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.045886 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046132 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046290 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046562 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046751 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046767 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046825 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046899 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046934 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046951 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.046972 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047127 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047218 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047261 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.064091 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.047280 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048080 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048108 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048423 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048544 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048708 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048706 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.048978 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.049509 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.049741 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.049749 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.049815 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.049978 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.049863 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.050092 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.050135 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.050265 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.050333 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.050340 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.050587 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.050599 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.050709 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.050759 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.050809 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.051066 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.051074 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.051134 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.051217 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.051622 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.051742 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.053731 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.054328 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.055789 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.055796 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.055808 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.055891 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.055975 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.056050 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.056212 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:06:53.5556313 +0000 UTC m=+21.194360643 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.056260 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.056355 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.056153 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.056445 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.056194 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.056501 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.056551 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.056760 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.056766 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.056782 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.056909 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.057221 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.057563 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.057595 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.057815 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.057954 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.058644 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.058735 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.058835 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.058849 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.058883 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.058895 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.059570 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.059592 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.059585 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.059601 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.059951 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.060010 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.060271 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.060314 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.060570 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.060999 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.061023 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.061163 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.061196 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.061469 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.061585 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.061652 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.061880 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.062025 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.062072 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.062100 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.062523 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.062585 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.062585 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.062617 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.062639 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.062999 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.063123 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.063279 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.063304 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.063512 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.063699 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.063715 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.063767 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.064275 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.064346 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.064619 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.064845 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.065118 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.065113 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.065145 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.065149 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.065448 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.065467 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.065943 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.066086 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.066177 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.066200 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.066554 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.066585 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.066690 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.066699 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.066786 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.066979 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.067103 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.067170 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.067211 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.067354 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.067604 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.068638 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.068774 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.068823 4857 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.068948 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.068967 4857 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.068977 4857 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.069550 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.069580 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.070203 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.070228 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.070301 4857 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.070415 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-22 00:06:53.570399481 +0000 UTC m=+21.209128734 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.070501 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.070625 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.070668 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.070793 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.070869 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.070935 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.070978 4857 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.071022 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.071086 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-22 00:06:53.571064131 +0000 UTC m=+21.209793394 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.071260 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.071273 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.071298 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.071372 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.071389 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.071666 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.071735 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.071811 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.072119 4857 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.072178 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.074614 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.076026 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.076076 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.076294 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.076545 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.076610 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.076906 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.076922 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.077281 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.077251 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.077151 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.077557 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.077640 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.077903 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.078025 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.078027 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.084128 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.085404 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.090502 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.090515 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.090694 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.090845 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.091165 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.091262 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.093408 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.093516 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.093560 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.093899 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.093927 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.093941 4857 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.094157 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-22 00:06:53.594135959 +0000 UTC m=+21.232865312 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.094267 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.094548 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.094572 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.094586 4857 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.094624 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-22 00:06:53.594612932 +0000 UTC m=+21.233342285 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.095092 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.095429 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.096334 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.097363 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.097612 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.098211 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.098946 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.100203 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.101372 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.104369 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.107383 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.107793 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.108569 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.110499 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.111565 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.113357 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.120876 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.121737 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.123104 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.124827 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.130172 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.131687 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.132458 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.135231 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.138669 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.143460 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.143645 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.143872 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.144189 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.146072 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.146656 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.147649 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.148220 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.148725 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.149675 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.150120 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.150971 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.151668 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.152207 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.153234 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.153648 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.154596 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.154808 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.155262 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.156443 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.157248 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.157825 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.158765 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.159212 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.160010 4857 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.160270 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.161869 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.162498 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.162986 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.165074 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.165636 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.166213 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.166793 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.168101 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.168933 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169343 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169427 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169473 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169510 4857 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169526 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169539 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169545 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169551 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169565 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169576 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169588 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169600 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169610 4857 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169622 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169633 4857 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169643 4857 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169654 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169664 4857 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169674 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169685 4857 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169695 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169706 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169717 4857 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169728 4857 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.169849 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170115 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170135 4857 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170170 4857 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170191 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170203 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170215 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170227 4857 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170238 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170250 4857 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170263 4857 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170274 4857 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170317 4857 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170327 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170336 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170346 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170355 4857 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170365 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170373 4857 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170413 4857 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170422 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170431 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170440 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170449 4857 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170458 4857 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170469 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170479 4857 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170488 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170497 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170505 4857 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170514 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170524 4857 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170533 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170542 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170550 4857 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170559 4857 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170568 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170577 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170586 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170595 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170604 4857 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170613 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170622 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170691 4857 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170710 4857 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170722 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170736 4857 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170748 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170759 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170771 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170782 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170793 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170804 4857 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170815 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170832 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170844 4857 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170857 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170869 4857 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170881 4857 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170887 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170892 4857 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170913 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170926 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170938 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170951 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170962 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170974 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170985 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.170997 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171010 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171022 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171049 4857 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171061 4857 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171072 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171083 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171093 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171105 4857 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171116 4857 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171126 4857 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171135 4857 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171145 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171156 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171166 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171176 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171186 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171196 4857 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171206 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171216 4857 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171227 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171236 4857 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171246 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171256 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171266 4857 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171275 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171305 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171314 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171325 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171335 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171346 4857 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171356 4857 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171366 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171377 4857 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171421 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171655 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171673 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171684 4857 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171694 4857 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171704 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171715 4857 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171725 4857 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171735 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171744 4857 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171753 4857 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171762 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171769 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171777 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171785 4857 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171809 4857 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171819 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171827 4857 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171968 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171977 4857 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171985 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.171994 4857 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172002 4857 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172010 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172018 4857 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172026 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172052 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172060 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172064 4857 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172163 4857 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172172 4857 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172181 4857 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172190 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172200 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172209 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172217 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172228 4857 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172237 4857 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172245 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172252 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172261 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172269 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172277 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172285 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172293 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172301 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172309 4857 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172318 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172325 4857 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172333 4857 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172341 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172351 4857 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172359 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172367 4857 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172376 4857 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172385 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172393 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172400 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172409 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172417 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172425 4857 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172432 4857 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172441 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172449 4857 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172456 4857 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172464 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172472 4857 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172480 4857 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172489 4857 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172498 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.172776 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.173786 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.174463 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.175738 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.176858 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.176896 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.177565 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.178664 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.179144 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.180020 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.180713 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.181175 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.185662 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.193440 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.201157 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.220844 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.221330 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.222811 4857 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253" exitCode=255 Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.222842 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253"} Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.222881 4857 scope.go:117] "RemoveContainer" containerID="43984627e32a20f5c61d14d0badf2782f948f3817107a051a8b61d0c8aaa1c2c" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.234015 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.234704 4857 scope.go:117] "RemoveContainer" containerID="190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253" Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.235115 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.236292 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.243224 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.250444 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.258574 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.271408 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.279169 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.315595 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.325547 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.331639 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 22 00:06:53 crc kubenswrapper[4857]: W0222 00:06:53.400943 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-4632d276e03c855f35c1abbbad7fe5847c7b3a2669bfdfcc0da1b616e63d388d WatchSource:0}: Error finding container 4632d276e03c855f35c1abbbad7fe5847c7b3a2669bfdfcc0da1b616e63d388d: Status 404 returned error can't find the container with id 4632d276e03c855f35c1abbbad7fe5847c7b3a2669bfdfcc0da1b616e63d388d Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.427714 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.442773 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.446203 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.448235 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43984627e32a20f5c61d14d0badf2782f948f3817107a051a8b61d0c8aaa1c2c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:47Z\\\",\\\"message\\\":\\\"W0222 00:06:36.341371 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0222 00:06:36.341784 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771718796 cert, and key in /tmp/serving-cert-1396102052/serving-signer.crt, /tmp/serving-cert-1396102052/serving-signer.key\\\\nI0222 00:06:36.533053 1 observer_polling.go:159] Starting file observer\\\\nW0222 00:06:36.535957 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0222 00:06:36.536095 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:36.536682 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1396102052/tls.crt::/tmp/serving-cert-1396102052/tls.key\\\\\\\"\\\\nF0222 00:06:46.991514 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.460594 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.472151 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.482987 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.495010 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.508863 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.518748 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.531439 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43984627e32a20f5c61d14d0badf2782f948f3817107a051a8b61d0c8aaa1c2c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:47Z\\\",\\\"message\\\":\\\"W0222 00:06:36.341371 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0222 00:06:36.341784 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771718796 cert, and key in /tmp/serving-cert-1396102052/serving-signer.crt, /tmp/serving-cert-1396102052/serving-signer.key\\\\nI0222 00:06:36.533053 1 observer_polling.go:159] Starting file observer\\\\nW0222 00:06:36.535957 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0222 00:06:36.536095 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:36.536682 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1396102052/tls.crt::/tmp/serving-cert-1396102052/tls.key\\\\\\\"\\\\nF0222 00:06:46.991514 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.549920 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.568941 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.574932 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.575027 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.575075 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:06:54.575054905 +0000 UTC m=+22.213784168 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.575107 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.575179 4857 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.575204 4857 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.575216 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-22 00:06:54.57520813 +0000 UTC m=+22.213937373 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.575247 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-22 00:06:54.575237991 +0000 UTC m=+22.213967244 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.583495 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.595094 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.603786 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.616074 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.631777 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.676155 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.676202 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.676317 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.676331 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.676342 4857 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.676383 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-22 00:06:54.676369627 +0000 UTC m=+22.315098880 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.676427 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.676436 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.676443 4857 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:06:53 crc kubenswrapper[4857]: E0222 00:06:53.676463 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-22 00:06:54.67645745 +0000 UTC m=+22.315186703 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.769645 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.784090 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.784409 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.798458 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.810261 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.819858 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.832317 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.840999 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.850261 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43984627e32a20f5c61d14d0badf2782f948f3817107a051a8b61d0c8aaa1c2c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:47Z\\\",\\\"message\\\":\\\"W0222 00:06:36.341371 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0222 00:06:36.341784 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771718796 cert, and key in /tmp/serving-cert-1396102052/serving-signer.crt, /tmp/serving-cert-1396102052/serving-signer.key\\\\nI0222 00:06:36.533053 1 observer_polling.go:159] Starting file observer\\\\nW0222 00:06:36.535957 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0222 00:06:36.536095 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:36.536682 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1396102052/tls.crt::/tmp/serving-cert-1396102052/tls.key\\\\\\\"\\\\nF0222 00:06:46.991514 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.860497 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.870755 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.878947 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.888239 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43984627e32a20f5c61d14d0badf2782f948f3817107a051a8b61d0c8aaa1c2c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:47Z\\\",\\\"message\\\":\\\"W0222 00:06:36.341371 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0222 00:06:36.341784 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771718796 cert, and key in /tmp/serving-cert-1396102052/serving-signer.crt, /tmp/serving-cert-1396102052/serving-signer.key\\\\nI0222 00:06:36.533053 1 observer_polling.go:159] Starting file observer\\\\nW0222 00:06:36.535957 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0222 00:06:36.536095 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:36.536682 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1396102052/tls.crt::/tmp/serving-cert-1396102052/tls.key\\\\\\\"\\\\nF0222 00:06:46.991514 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.896135 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.910593 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.919652 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.931924 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.942299 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.955782 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:53 crc kubenswrapper[4857]: I0222 00:06:53.972126 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.000674 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 01:05:51.877194521 +0000 UTC Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.077107 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:06:54 crc kubenswrapper[4857]: E0222 00:06:54.077229 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.226626 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb"} Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.226670 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20"} Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.226680 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"fd2d4aa783272b4088b0d91cb00b09697575a11646d6ef89a9cb4b47fb0c69cc"} Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.228431 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.230756 4857 scope.go:117] "RemoveContainer" containerID="190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253" Feb 22 00:06:54 crc kubenswrapper[4857]: E0222 00:06:54.231096 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.231665 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6"} Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.231689 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"4632d276e03c855f35c1abbbad7fe5847c7b3a2669bfdfcc0da1b616e63d388d"} Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.232827 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"9c674a4a1d5eb92042ab5f237bf489a26fb199f0dcca188f205464e730234023"} Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.239938 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.253812 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.271835 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.286606 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.300417 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.315135 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.327691 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://43984627e32a20f5c61d14d0badf2782f948f3817107a051a8b61d0c8aaa1c2c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:47Z\\\",\\\"message\\\":\\\"W0222 00:06:36.341371 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0222 00:06:36.341784 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771718796 cert, and key in /tmp/serving-cert-1396102052/serving-signer.crt, /tmp/serving-cert-1396102052/serving-signer.key\\\\nI0222 00:06:36.533053 1 observer_polling.go:159] Starting file observer\\\\nW0222 00:06:36.535957 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0222 00:06:36.536095 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:36.536682 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1396102052/tls.crt::/tmp/serving-cert-1396102052/tls.key\\\\\\\"\\\\nF0222 00:06:46.991514 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.339114 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.349576 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.359968 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.373381 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.390984 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.406852 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.418002 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.428815 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.439571 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.450581 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.461706 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.589422 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.589490 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.589512 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:06:54 crc kubenswrapper[4857]: E0222 00:06:54.589565 4857 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 22 00:06:54 crc kubenswrapper[4857]: E0222 00:06:54.589574 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:06:56.5895461 +0000 UTC m=+24.228275353 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:06:54 crc kubenswrapper[4857]: E0222 00:06:54.589604 4857 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 22 00:06:54 crc kubenswrapper[4857]: E0222 00:06:54.589609 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-22 00:06:56.589601572 +0000 UTC m=+24.228330825 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 22 00:06:54 crc kubenswrapper[4857]: E0222 00:06:54.589656 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-22 00:06:56.589648853 +0000 UTC m=+24.228378106 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.690761 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:06:54 crc kubenswrapper[4857]: I0222 00:06:54.690819 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:06:54 crc kubenswrapper[4857]: E0222 00:06:54.690934 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 22 00:06:54 crc kubenswrapper[4857]: E0222 00:06:54.690941 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 22 00:06:54 crc kubenswrapper[4857]: E0222 00:06:54.690995 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 22 00:06:54 crc kubenswrapper[4857]: E0222 00:06:54.691010 4857 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:06:54 crc kubenswrapper[4857]: E0222 00:06:54.690949 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 22 00:06:54 crc kubenswrapper[4857]: E0222 00:06:54.691072 4857 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:06:54 crc kubenswrapper[4857]: E0222 00:06:54.691078 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-22 00:06:56.691061458 +0000 UTC m=+24.329790711 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:06:54 crc kubenswrapper[4857]: E0222 00:06:54.691127 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-22 00:06:56.691104699 +0000 UTC m=+24.329833952 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:06:55 crc kubenswrapper[4857]: I0222 00:06:55.002114 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 18:26:45.993620902 +0000 UTC Feb 22 00:06:55 crc kubenswrapper[4857]: I0222 00:06:55.076783 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:06:55 crc kubenswrapper[4857]: E0222 00:06:55.076894 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:06:55 crc kubenswrapper[4857]: I0222 00:06:55.076786 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:06:55 crc kubenswrapper[4857]: E0222 00:06:55.077027 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:06:55 crc kubenswrapper[4857]: I0222 00:06:55.237267 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a"} Feb 22 00:06:55 crc kubenswrapper[4857]: I0222 00:06:55.255370 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:55Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:55 crc kubenswrapper[4857]: I0222 00:06:55.266804 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:55Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:55 crc kubenswrapper[4857]: I0222 00:06:55.276400 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:55Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:55 crc kubenswrapper[4857]: I0222 00:06:55.285461 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:55Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:55 crc kubenswrapper[4857]: I0222 00:06:55.295672 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:55Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:55 crc kubenswrapper[4857]: I0222 00:06:55.306133 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:55Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:55 crc kubenswrapper[4857]: I0222 00:06:55.318837 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:55Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:55 crc kubenswrapper[4857]: I0222 00:06:55.333838 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:55Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:55 crc kubenswrapper[4857]: I0222 00:06:55.349493 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:55Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.003012 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 06:01:56.665747604 +0000 UTC Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.076846 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:06:56 crc kubenswrapper[4857]: E0222 00:06:56.077164 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.633921 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:06:56 crc kubenswrapper[4857]: E0222 00:06:56.634074 4857 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.633991 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:06:56 crc kubenswrapper[4857]: E0222 00:06:56.634166 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:07:00.634113336 +0000 UTC m=+28.272842589 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:06:56 crc kubenswrapper[4857]: E0222 00:06:56.634222 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:00.634213468 +0000 UTC m=+28.272942711 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.634320 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:06:56 crc kubenswrapper[4857]: E0222 00:06:56.634458 4857 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 22 00:06:56 crc kubenswrapper[4857]: E0222 00:06:56.634512 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:00.634503787 +0000 UTC m=+28.273233040 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.652145 4857 csr.go:261] certificate signing request csr-kktsh is approved, waiting to be issued Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.669240 4857 csr.go:257] certificate signing request csr-kktsh is issued Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.735224 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.735280 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:06:56 crc kubenswrapper[4857]: E0222 00:06:56.735387 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 22 00:06:56 crc kubenswrapper[4857]: E0222 00:06:56.735389 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 22 00:06:56 crc kubenswrapper[4857]: E0222 00:06:56.735426 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 22 00:06:56 crc kubenswrapper[4857]: E0222 00:06:56.735437 4857 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:06:56 crc kubenswrapper[4857]: E0222 00:06:56.735496 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:00.735481469 +0000 UTC m=+28.374210722 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:06:56 crc kubenswrapper[4857]: E0222 00:06:56.735403 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 22 00:06:56 crc kubenswrapper[4857]: E0222 00:06:56.735516 4857 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:06:56 crc kubenswrapper[4857]: E0222 00:06:56.735557 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:00.735544571 +0000 UTC m=+28.374273824 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.744680 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-2vqwv"] Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.745026 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-lz8x8"] Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.745134 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-2vqwv" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.745273 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-lz8x8" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.747613 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.747778 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.747826 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.748071 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.748088 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.748076 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.748980 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.765984 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.780173 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.792902 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.804951 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.819028 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.836258 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbmn6\" (UniqueName: \"kubernetes.io/projected/35f87b2b-119f-455d-9a79-39dd15ac2559-kube-api-access-bbmn6\") pod \"node-resolver-2vqwv\" (UID: \"35f87b2b-119f-455d-9a79-39dd15ac2559\") " pod="openshift-dns/node-resolver-2vqwv" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.836316 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d9642d27-573f-4a44-b37c-a636dc22439f-serviceca\") pod \"node-ca-lz8x8\" (UID: \"d9642d27-573f-4a44-b37c-a636dc22439f\") " pod="openshift-image-registry/node-ca-lz8x8" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.836343 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r98zd\" (UniqueName: \"kubernetes.io/projected/d9642d27-573f-4a44-b37c-a636dc22439f-kube-api-access-r98zd\") pod \"node-ca-lz8x8\" (UID: \"d9642d27-573f-4a44-b37c-a636dc22439f\") " pod="openshift-image-registry/node-ca-lz8x8" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.836387 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/35f87b2b-119f-455d-9a79-39dd15ac2559-hosts-file\") pod \"node-resolver-2vqwv\" (UID: \"35f87b2b-119f-455d-9a79-39dd15ac2559\") " pod="openshift-dns/node-resolver-2vqwv" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.836431 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d9642d27-573f-4a44-b37c-a636dc22439f-host\") pod \"node-ca-lz8x8\" (UID: \"d9642d27-573f-4a44-b37c-a636dc22439f\") " pod="openshift-image-registry/node-ca-lz8x8" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.842393 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.866314 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.882118 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.897446 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.914948 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.937147 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d9642d27-573f-4a44-b37c-a636dc22439f-host\") pod \"node-ca-lz8x8\" (UID: \"d9642d27-573f-4a44-b37c-a636dc22439f\") " pod="openshift-image-registry/node-ca-lz8x8" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.937228 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbmn6\" (UniqueName: \"kubernetes.io/projected/35f87b2b-119f-455d-9a79-39dd15ac2559-kube-api-access-bbmn6\") pod \"node-resolver-2vqwv\" (UID: \"35f87b2b-119f-455d-9a79-39dd15ac2559\") " pod="openshift-dns/node-resolver-2vqwv" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.937249 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d9642d27-573f-4a44-b37c-a636dc22439f-serviceca\") pod \"node-ca-lz8x8\" (UID: \"d9642d27-573f-4a44-b37c-a636dc22439f\") " pod="openshift-image-registry/node-ca-lz8x8" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.937269 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r98zd\" (UniqueName: \"kubernetes.io/projected/d9642d27-573f-4a44-b37c-a636dc22439f-kube-api-access-r98zd\") pod \"node-ca-lz8x8\" (UID: \"d9642d27-573f-4a44-b37c-a636dc22439f\") " pod="openshift-image-registry/node-ca-lz8x8" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.937290 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d9642d27-573f-4a44-b37c-a636dc22439f-host\") pod \"node-ca-lz8x8\" (UID: \"d9642d27-573f-4a44-b37c-a636dc22439f\") " pod="openshift-image-registry/node-ca-lz8x8" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.937303 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/35f87b2b-119f-455d-9a79-39dd15ac2559-hosts-file\") pod \"node-resolver-2vqwv\" (UID: \"35f87b2b-119f-455d-9a79-39dd15ac2559\") " pod="openshift-dns/node-resolver-2vqwv" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.937420 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/35f87b2b-119f-455d-9a79-39dd15ac2559-hosts-file\") pod \"node-resolver-2vqwv\" (UID: \"35f87b2b-119f-455d-9a79-39dd15ac2559\") " pod="openshift-dns/node-resolver-2vqwv" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.938258 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d9642d27-573f-4a44-b37c-a636dc22439f-serviceca\") pod \"node-ca-lz8x8\" (UID: \"d9642d27-573f-4a44-b37c-a636dc22439f\") " pod="openshift-image-registry/node-ca-lz8x8" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.943501 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.963546 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.970712 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r98zd\" (UniqueName: \"kubernetes.io/projected/d9642d27-573f-4a44-b37c-a636dc22439f-kube-api-access-r98zd\") pod \"node-ca-lz8x8\" (UID: \"d9642d27-573f-4a44-b37c-a636dc22439f\") " pod="openshift-image-registry/node-ca-lz8x8" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.974864 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbmn6\" (UniqueName: \"kubernetes.io/projected/35f87b2b-119f-455d-9a79-39dd15ac2559-kube-api-access-bbmn6\") pod \"node-resolver-2vqwv\" (UID: \"35f87b2b-119f-455d-9a79-39dd15ac2559\") " pod="openshift-dns/node-resolver-2vqwv" Feb 22 00:06:56 crc kubenswrapper[4857]: I0222 00:06:56.989805 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.003268 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 23:41:13.743020481 +0000 UTC Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.003741 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.024493 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.041734 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.052725 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.056651 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-lz8x8" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.063114 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-2vqwv" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.070927 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.077230 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.077252 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:06:57 crc kubenswrapper[4857]: E0222 00:06:57.077353 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:06:57 crc kubenswrapper[4857]: E0222 00:06:57.077482 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.086650 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.104800 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.116968 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.146616 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-4lmlt"] Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.146996 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6rc8w"] Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.147294 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.147962 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-zd2xg"] Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.148128 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.148931 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-gw6k5"] Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.149204 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.149207 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.149725 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.149789 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.150029 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.152677 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.152941 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.153269 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.153413 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.154064 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.154100 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.164155 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.164249 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.165653 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.165817 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.165982 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.166154 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.166342 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.166481 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.169291 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.170529 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.175930 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.188745 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.199304 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.213680 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.226879 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240091 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-run-systemd\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240143 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9a0db241-7fc5-4d71-a8cc-534f4d303883-env-overrides\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240168 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28jk9\" (UniqueName: \"kubernetes.io/projected/fec51cf7-f11e-4829-bcd2-08ac17cdce91-kube-api-access-28jk9\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240209 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-os-release\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240235 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2ed2fde8-1449-491b-ae21-2fe11ab008ae-mcd-auth-proxy-config\") pod \"machine-config-daemon-gw6k5\" (UID: \"2ed2fde8-1449-491b-ae21-2fe11ab008ae\") " pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240255 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9a0db241-7fc5-4d71-a8cc-534f4d303883-ovnkube-config\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240287 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-hostroot\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240307 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-multus-conf-dir\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240326 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-var-lib-openvswitch\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240347 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-cnibin\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240374 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-multus-daemon-config\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240396 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-log-socket\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240418 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-host-var-lib-cni-bin\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240439 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-host-run-multus-certs\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240459 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-kubelet\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240517 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2ed2fde8-1449-491b-ae21-2fe11ab008ae-rootfs\") pod \"machine-config-daemon-gw6k5\" (UID: \"2ed2fde8-1449-491b-ae21-2fe11ab008ae\") " pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240540 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-multus-cni-dir\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240562 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-node-log\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240584 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fw59\" (UniqueName: \"kubernetes.io/projected/9a0db241-7fc5-4d71-a8cc-534f4d303883-kube-api-access-7fw59\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240606 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-run-ovn-kubernetes\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240629 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240661 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9a0db241-7fc5-4d71-a8cc-534f4d303883-ovnkube-script-lib\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240728 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240857 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-host-run-netns\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240932 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-slash\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.240993 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-run-ovn\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241028 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fec51cf7-f11e-4829-bcd2-08ac17cdce91-system-cni-dir\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241091 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fec51cf7-f11e-4829-bcd2-08ac17cdce91-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241131 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-run-openvswitch\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241162 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-cni-bin\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241185 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-cni-netd\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241237 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-host-run-k8s-cni-cncf-io\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241264 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-host-var-lib-kubelet\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241288 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-systemd-units\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241328 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-etc-openvswitch\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241355 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9a0db241-7fc5-4d71-a8cc-534f4d303883-ovn-node-metrics-cert\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241381 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fec51cf7-f11e-4829-bcd2-08ac17cdce91-cnibin\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241404 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fec51cf7-f11e-4829-bcd2-08ac17cdce91-os-release\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241484 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2kwn\" (UniqueName: \"kubernetes.io/projected/2ed2fde8-1449-491b-ae21-2fe11ab008ae-kube-api-access-m2kwn\") pod \"machine-config-daemon-gw6k5\" (UID: \"2ed2fde8-1449-491b-ae21-2fe11ab008ae\") " pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241527 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-etc-kubernetes\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241549 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-cni-binary-copy\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241582 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-host-var-lib-cni-multus\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241604 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-run-netns\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241627 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fec51cf7-f11e-4829-bcd2-08ac17cdce91-cni-binary-copy\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241695 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2ed2fde8-1449-491b-ae21-2fe11ab008ae-proxy-tls\") pod \"machine-config-daemon-gw6k5\" (UID: \"2ed2fde8-1449-491b-ae21-2fe11ab008ae\") " pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241723 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-system-cni-dir\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241755 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-multus-socket-dir-parent\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241782 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fec51cf7-f11e-4829-bcd2-08ac17cdce91-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.241817 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwrkb\" (UniqueName: \"kubernetes.io/projected/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-kube-api-access-kwrkb\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.244681 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-2vqwv" event={"ID":"35f87b2b-119f-455d-9a79-39dd15ac2559","Type":"ContainerStarted","Data":"a126cef2597692a04bf10674956b9378fbd8166eb4743d1f00e1607d7ce0bbe3"} Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.246055 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-lz8x8" event={"ID":"d9642d27-573f-4a44-b37c-a636dc22439f","Type":"ContainerStarted","Data":"4c9148002ae57c609035e90286fc89f371f53c843aae6f2463405429879a147b"} Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.262330 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.285150 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.299448 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.316729 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.330441 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342273 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-cni-bin\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342323 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-cni-netd\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342354 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-host-run-k8s-cni-cncf-io\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342376 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-host-var-lib-kubelet\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342397 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-systemd-units\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342418 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-etc-openvswitch\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342431 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-host-run-k8s-cni-cncf-io\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342431 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-cni-bin\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342441 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9a0db241-7fc5-4d71-a8cc-534f4d303883-ovn-node-metrics-cert\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342501 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-etc-openvswitch\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342517 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fec51cf7-f11e-4829-bcd2-08ac17cdce91-cnibin\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342524 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-host-var-lib-kubelet\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342571 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/fec51cf7-f11e-4829-bcd2-08ac17cdce91-cnibin\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342536 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fec51cf7-f11e-4829-bcd2-08ac17cdce91-os-release\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342443 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-systemd-units\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342413 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-cni-netd\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342783 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/fec51cf7-f11e-4829-bcd2-08ac17cdce91-os-release\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342638 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2kwn\" (UniqueName: \"kubernetes.io/projected/2ed2fde8-1449-491b-ae21-2fe11ab008ae-kube-api-access-m2kwn\") pod \"machine-config-daemon-gw6k5\" (UID: \"2ed2fde8-1449-491b-ae21-2fe11ab008ae\") " pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342817 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-etc-kubernetes\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342833 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-cni-binary-copy\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342850 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-host-var-lib-cni-multus\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342865 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-run-netns\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342880 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fec51cf7-f11e-4829-bcd2-08ac17cdce91-cni-binary-copy\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342883 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-etc-kubernetes\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342901 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2ed2fde8-1449-491b-ae21-2fe11ab008ae-proxy-tls\") pod \"machine-config-daemon-gw6k5\" (UID: \"2ed2fde8-1449-491b-ae21-2fe11ab008ae\") " pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342919 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-system-cni-dir\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342934 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-run-netns\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342937 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-multus-socket-dir-parent\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342966 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-multus-socket-dir-parent\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.342983 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fec51cf7-f11e-4829-bcd2-08ac17cdce91-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343004 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwrkb\" (UniqueName: \"kubernetes.io/projected/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-kube-api-access-kwrkb\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343007 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-host-var-lib-cni-multus\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343031 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-run-systemd\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343066 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9a0db241-7fc5-4d71-a8cc-534f4d303883-env-overrides\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343083 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28jk9\" (UniqueName: \"kubernetes.io/projected/fec51cf7-f11e-4829-bcd2-08ac17cdce91-kube-api-access-28jk9\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343101 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-os-release\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343117 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2ed2fde8-1449-491b-ae21-2fe11ab008ae-mcd-auth-proxy-config\") pod \"machine-config-daemon-gw6k5\" (UID: \"2ed2fde8-1449-491b-ae21-2fe11ab008ae\") " pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343118 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-system-cni-dir\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343143 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9a0db241-7fc5-4d71-a8cc-534f4d303883-ovnkube-config\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343166 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-hostroot\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343182 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-multus-conf-dir\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343196 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-var-lib-openvswitch\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343213 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-cnibin\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343229 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-multus-daemon-config\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343246 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-log-socket\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343264 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-host-var-lib-cni-bin\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343267 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-hostroot\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343307 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-multus-conf-dir\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343310 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-host-run-multus-certs\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343335 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-var-lib-openvswitch\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343353 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-run-systemd\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343370 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-cnibin\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343281 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-host-run-multus-certs\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343399 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-kubelet\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343418 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2ed2fde8-1449-491b-ae21-2fe11ab008ae-rootfs\") pod \"machine-config-daemon-gw6k5\" (UID: \"2ed2fde8-1449-491b-ae21-2fe11ab008ae\") " pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343433 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-multus-cni-dir\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343446 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-node-log\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343462 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fw59\" (UniqueName: \"kubernetes.io/projected/9a0db241-7fc5-4d71-a8cc-534f4d303883-kube-api-access-7fw59\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343477 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-run-ovn-kubernetes\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343493 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343527 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9a0db241-7fc5-4d71-a8cc-534f4d303883-ovnkube-script-lib\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343536 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-log-socket\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343563 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-host-run-netns\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343580 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-slash\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343596 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-run-ovn\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343613 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fec51cf7-f11e-4829-bcd2-08ac17cdce91-system-cni-dir\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343629 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fec51cf7-f11e-4829-bcd2-08ac17cdce91-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343646 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-run-openvswitch\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343690 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-run-openvswitch\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343713 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-kubelet\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343733 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/2ed2fde8-1449-491b-ae21-2fe11ab008ae-rootfs\") pod \"machine-config-daemon-gw6k5\" (UID: \"2ed2fde8-1449-491b-ae21-2fe11ab008ae\") " pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343767 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-multus-cni-dir\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343788 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-node-log\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343789 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-cni-binary-copy\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343843 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-host-var-lib-cni-bin\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.344089 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-run-ovn-kubernetes\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.344120 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.344191 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-slash\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.344219 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-run-ovn\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.344217 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/fec51cf7-f11e-4829-bcd2-08ac17cdce91-cni-binary-copy\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.344252 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/fec51cf7-f11e-4829-bcd2-08ac17cdce91-system-cni-dir\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.344451 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9a0db241-7fc5-4d71-a8cc-534f4d303883-ovnkube-config\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.344461 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/fec51cf7-f11e-4829-bcd2-08ac17cdce91-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.344486 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/2ed2fde8-1449-491b-ae21-2fe11ab008ae-mcd-auth-proxy-config\") pod \"machine-config-daemon-gw6k5\" (UID: \"2ed2fde8-1449-491b-ae21-2fe11ab008ae\") " pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.343357 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-os-release\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.344465 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-host-run-netns\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.344616 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-multus-daemon-config\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.344735 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9a0db241-7fc5-4d71-a8cc-534f4d303883-ovnkube-script-lib\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.344781 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/fec51cf7-f11e-4829-bcd2-08ac17cdce91-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.344788 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9a0db241-7fc5-4d71-a8cc-534f4d303883-env-overrides\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.350489 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/2ed2fde8-1449-491b-ae21-2fe11ab008ae-proxy-tls\") pod \"machine-config-daemon-gw6k5\" (UID: \"2ed2fde8-1449-491b-ae21-2fe11ab008ae\") " pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.353720 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9a0db241-7fc5-4d71-a8cc-534f4d303883-ovn-node-metrics-cert\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.362908 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.370931 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fw59\" (UniqueName: \"kubernetes.io/projected/9a0db241-7fc5-4d71-a8cc-534f4d303883-kube-api-access-7fw59\") pod \"ovnkube-node-6rc8w\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.373637 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwrkb\" (UniqueName: \"kubernetes.io/projected/9e857d8a-7289-4352-9de9-1b5d0bd21e8f-kube-api-access-kwrkb\") pod \"multus-4lmlt\" (UID: \"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\") " pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.373781 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2kwn\" (UniqueName: \"kubernetes.io/projected/2ed2fde8-1449-491b-ae21-2fe11ab008ae-kube-api-access-m2kwn\") pod \"machine-config-daemon-gw6k5\" (UID: \"2ed2fde8-1449-491b-ae21-2fe11ab008ae\") " pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.376157 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28jk9\" (UniqueName: \"kubernetes.io/projected/fec51cf7-f11e-4829-bcd2-08ac17cdce91-kube-api-access-28jk9\") pod \"multus-additional-cni-plugins-zd2xg\" (UID: \"fec51cf7-f11e-4829-bcd2-08ac17cdce91\") " pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.390398 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.417990 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.438104 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.466599 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.485435 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4lmlt" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.490811 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.494372 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.502661 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:06:57 crc kubenswrapper[4857]: W0222 00:06:57.502990 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e857d8a_7289_4352_9de9_1b5d0bd21e8f.slice/crio-431842afc58cf4cf47b14b3ec5d9ace609008416e68ba93b9acfdb418fa28beb WatchSource:0}: Error finding container 431842afc58cf4cf47b14b3ec5d9ace609008416e68ba93b9acfdb418fa28beb: Status 404 returned error can't find the container with id 431842afc58cf4cf47b14b3ec5d9ace609008416e68ba93b9acfdb418fa28beb Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.508804 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.578561 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.617244 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.650055 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.664477 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.670381 4857 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-22 00:01:56 +0000 UTC, rotation deadline is 2026-11-17 05:35:38.983209853 +0000 UTC Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.670423 4857 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6437h28m41.312788419s for next certificate rotation Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.675130 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.688428 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.704391 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.724299 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.753022 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:57 crc kubenswrapper[4857]: I0222 00:06:57.766017 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.004022 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 14:34:22.560702591 +0000 UTC Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.077523 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:06:58 crc kubenswrapper[4857]: E0222 00:06:58.077644 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.250632 4857 generic.go:334] "Generic (PLEG): container finished" podID="fec51cf7-f11e-4829-bcd2-08ac17cdce91" containerID="42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f" exitCode=0 Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.250684 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" event={"ID":"fec51cf7-f11e-4829-bcd2-08ac17cdce91","Type":"ContainerDied","Data":"42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f"} Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.250734 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" event={"ID":"fec51cf7-f11e-4829-bcd2-08ac17cdce91","Type":"ContainerStarted","Data":"a0271cc800832e12207a2271391d0ab219172c6b2eae0ab6ee5bf570da159563"} Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.251810 4857 generic.go:334] "Generic (PLEG): container finished" podID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerID="f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b" exitCode=0 Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.251855 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerDied","Data":"f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b"} Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.251881 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerStarted","Data":"01681e4530f602e5e717500243924331ffb7efdd3b4aca046931512101cb8c6f"} Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.253714 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-lz8x8" event={"ID":"d9642d27-573f-4a44-b37c-a636dc22439f","Type":"ContainerStarted","Data":"be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79"} Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.256716 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" event={"ID":"2ed2fde8-1449-491b-ae21-2fe11ab008ae","Type":"ContainerStarted","Data":"ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8"} Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.256765 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" event={"ID":"2ed2fde8-1449-491b-ae21-2fe11ab008ae","Type":"ContainerStarted","Data":"bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7"} Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.256780 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" event={"ID":"2ed2fde8-1449-491b-ae21-2fe11ab008ae","Type":"ContainerStarted","Data":"40e0e7e5939e8ef108f69758345f2010fff94ac1324693454e0c4de0ac1d347c"} Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.261128 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4lmlt" event={"ID":"9e857d8a-7289-4352-9de9-1b5d0bd21e8f","Type":"ContainerStarted","Data":"f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15"} Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.261164 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4lmlt" event={"ID":"9e857d8a-7289-4352-9de9-1b5d0bd21e8f","Type":"ContainerStarted","Data":"431842afc58cf4cf47b14b3ec5d9ace609008416e68ba93b9acfdb418fa28beb"} Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.263086 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-2vqwv" event={"ID":"35f87b2b-119f-455d-9a79-39dd15ac2559","Type":"ContainerStarted","Data":"955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd"} Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.271576 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.290545 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.302595 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.317495 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.333174 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.348281 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.359284 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.372775 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.396947 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.409718 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.423786 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.437071 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.457315 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.475203 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.486535 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.501347 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.513175 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.532314 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.549131 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.563796 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.584110 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.604404 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.622716 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.633828 4857 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.639345 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.639387 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.639398 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.639502 4857 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.639356 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.649707 4857 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.650427 4857 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.651981 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.652027 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.652051 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.652071 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.652085 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:06:58Z","lastTransitionTime":"2026-02-22T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.662234 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.680412 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: E0222 00:06:58.685431 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.689473 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.689507 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.689516 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.689530 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.689541 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:06:58Z","lastTransitionTime":"2026-02-22T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.708866 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: E0222 00:06:58.762439 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.766155 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.768723 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.768899 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.768990 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.769104 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.769193 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:06:58Z","lastTransitionTime":"2026-02-22T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.778002 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: E0222 00:06:58.789201 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.795725 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.795789 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.795798 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.795814 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.795839 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:06:58Z","lastTransitionTime":"2026-02-22T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.796296 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: E0222 00:06:58.811790 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.816582 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.816621 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.816630 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.816645 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.816656 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:06:58Z","lastTransitionTime":"2026-02-22T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:06:58 crc kubenswrapper[4857]: E0222 00:06:58.828726 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:58 crc kubenswrapper[4857]: E0222 00:06:58.828847 4857 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.830818 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.830866 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.830881 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.830900 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.830913 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:06:58Z","lastTransitionTime":"2026-02-22T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.933159 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.933561 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.933573 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.933588 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:06:58 crc kubenswrapper[4857]: I0222 00:06:58.933599 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:06:58Z","lastTransitionTime":"2026-02-22T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.004515 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 16:18:41.36920506 +0000 UTC Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.038114 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.038157 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.038169 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.038188 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.038201 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:06:59Z","lastTransitionTime":"2026-02-22T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.076934 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.076946 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:06:59 crc kubenswrapper[4857]: E0222 00:06:59.077090 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:06:59 crc kubenswrapper[4857]: E0222 00:06:59.077116 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.140455 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.140490 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.140499 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.140514 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.140523 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:06:59Z","lastTransitionTime":"2026-02-22T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.243216 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.243272 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.243285 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.243302 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.243316 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:06:59Z","lastTransitionTime":"2026-02-22T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.268241 4857 generic.go:334] "Generic (PLEG): container finished" podID="fec51cf7-f11e-4829-bcd2-08ac17cdce91" containerID="c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132" exitCode=0 Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.268327 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" event={"ID":"fec51cf7-f11e-4829-bcd2-08ac17cdce91","Type":"ContainerDied","Data":"c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132"} Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.272336 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerStarted","Data":"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd"} Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.272375 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerStarted","Data":"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93"} Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.272388 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerStarted","Data":"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9"} Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.272399 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerStarted","Data":"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770"} Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.272413 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerStarted","Data":"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b"} Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.303023 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:59Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.320700 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:59Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.333918 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:59Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.345323 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.345369 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.345380 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.345395 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.345405 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:06:59Z","lastTransitionTime":"2026-02-22T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.348377 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:59Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.362965 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:59Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.379606 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:59Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.402294 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:59Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.417104 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:59Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.433540 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:59Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.448620 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.448652 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.448661 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.448677 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.448689 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:06:59Z","lastTransitionTime":"2026-02-22T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.450624 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:59Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.464380 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:59Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.484083 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:59Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.506773 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:59Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.519675 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:59Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.550084 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:06:59Z is after 2025-08-24T17:21:41Z" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.552731 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.552766 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.552776 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.552792 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.552802 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:06:59Z","lastTransitionTime":"2026-02-22T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.655202 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.655237 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.655245 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.655258 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.655269 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:06:59Z","lastTransitionTime":"2026-02-22T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.759150 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.759199 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.759213 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.759238 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.759251 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:06:59Z","lastTransitionTime":"2026-02-22T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.862233 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.862276 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.862287 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.862304 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.862316 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:06:59Z","lastTransitionTime":"2026-02-22T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.964568 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.964601 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.964608 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.964622 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:06:59 crc kubenswrapper[4857]: I0222 00:06:59.964630 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:06:59Z","lastTransitionTime":"2026-02-22T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.004813 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 05:48:07.876289365 +0000 UTC Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.066793 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.066836 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.066850 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.066867 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.066879 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:00Z","lastTransitionTime":"2026-02-22T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.076587 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:00 crc kubenswrapper[4857]: E0222 00:07:00.076765 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.169419 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.169479 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.169496 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.169526 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.169542 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:00Z","lastTransitionTime":"2026-02-22T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.272203 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.272247 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.272256 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.272274 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.272284 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:00Z","lastTransitionTime":"2026-02-22T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.276708 4857 generic.go:334] "Generic (PLEG): container finished" podID="fec51cf7-f11e-4829-bcd2-08ac17cdce91" containerID="2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d" exitCode=0 Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.276785 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" event={"ID":"fec51cf7-f11e-4829-bcd2-08ac17cdce91","Type":"ContainerDied","Data":"2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d"} Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.280917 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerStarted","Data":"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2"} Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.290076 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.323872 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.335439 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.349906 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.364885 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.376993 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.377084 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.377124 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.377134 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.377154 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.377168 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:00Z","lastTransitionTime":"2026-02-22T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.391359 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.408464 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.429561 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.445363 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.460144 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.474991 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.479806 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.479836 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.479847 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.479867 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.479880 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:00Z","lastTransitionTime":"2026-02-22T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.485569 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.496436 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.511059 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.583410 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.583461 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.583470 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.583491 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.583501 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:00Z","lastTransitionTime":"2026-02-22T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.679427 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.679565 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.679602 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:00 crc kubenswrapper[4857]: E0222 00:07:00.679757 4857 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 22 00:07:00 crc kubenswrapper[4857]: E0222 00:07:00.679775 4857 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 22 00:07:00 crc kubenswrapper[4857]: E0222 00:07:00.679848 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:08.679825499 +0000 UTC m=+36.318554752 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 22 00:07:00 crc kubenswrapper[4857]: E0222 00:07:00.679875 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:08.67986689 +0000 UTC m=+36.318596143 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 22 00:07:00 crc kubenswrapper[4857]: E0222 00:07:00.679900 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:07:08.679887241 +0000 UTC m=+36.318616504 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.686155 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.686194 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.686203 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.686220 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.686231 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:00Z","lastTransitionTime":"2026-02-22T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.780868 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.780905 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:00 crc kubenswrapper[4857]: E0222 00:07:00.781015 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 22 00:07:00 crc kubenswrapper[4857]: E0222 00:07:00.781029 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 22 00:07:00 crc kubenswrapper[4857]: E0222 00:07:00.781054 4857 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:07:00 crc kubenswrapper[4857]: E0222 00:07:00.781102 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:08.78108968 +0000 UTC m=+36.419818933 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:07:00 crc kubenswrapper[4857]: E0222 00:07:00.781204 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 22 00:07:00 crc kubenswrapper[4857]: E0222 00:07:00.781265 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 22 00:07:00 crc kubenswrapper[4857]: E0222 00:07:00.781289 4857 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:07:00 crc kubenswrapper[4857]: E0222 00:07:00.781407 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:08.781369998 +0000 UTC m=+36.420099381 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.788435 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.788505 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.788516 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.788533 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.788544 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:00Z","lastTransitionTime":"2026-02-22T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.891623 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.891662 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.891670 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.891685 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.891694 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:00Z","lastTransitionTime":"2026-02-22T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.993375 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.993410 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.993420 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.993445 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:00 crc kubenswrapper[4857]: I0222 00:07:00.993455 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:00Z","lastTransitionTime":"2026-02-22T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.004956 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 03:59:36.90239688 +0000 UTC Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.079246 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.079297 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:01 crc kubenswrapper[4857]: E0222 00:07:01.079860 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:01 crc kubenswrapper[4857]: E0222 00:07:01.080065 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.096313 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.096347 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.096357 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.096371 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.096381 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:01Z","lastTransitionTime":"2026-02-22T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.199067 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.199105 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.199115 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.199132 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.199141 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:01Z","lastTransitionTime":"2026-02-22T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.288746 4857 generic.go:334] "Generic (PLEG): container finished" podID="fec51cf7-f11e-4829-bcd2-08ac17cdce91" containerID="22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd" exitCode=0 Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.288795 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" event={"ID":"fec51cf7-f11e-4829-bcd2-08ac17cdce91","Type":"ContainerDied","Data":"22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd"} Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.302305 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.302402 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.302419 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.302436 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.302448 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:01Z","lastTransitionTime":"2026-02-22T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.304554 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:01Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.339216 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:01Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.352940 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:01Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.363960 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.364742 4857 scope.go:117] "RemoveContainer" containerID="190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253" Feb 22 00:07:01 crc kubenswrapper[4857]: E0222 00:07:01.364918 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.367407 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:01Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.382987 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:01Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.393398 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:01Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.408264 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.408294 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.408303 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.408316 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.408326 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:01Z","lastTransitionTime":"2026-02-22T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.408475 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:01Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.428129 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:01Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.438720 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:01Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.450350 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:01Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.464699 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:01Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.490282 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:01Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.522342 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.522405 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.522424 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.522488 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.522509 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:01Z","lastTransitionTime":"2026-02-22T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.524492 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:01Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.537308 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:01Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.552800 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:01Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.625563 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.625621 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.625637 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.625661 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.625678 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:01Z","lastTransitionTime":"2026-02-22T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.727617 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.727654 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.727663 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.727679 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.727691 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:01Z","lastTransitionTime":"2026-02-22T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.830249 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.830275 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.830283 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.830296 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.830324 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:01Z","lastTransitionTime":"2026-02-22T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.932642 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.932678 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.932686 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.932700 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:01 crc kubenswrapper[4857]: I0222 00:07:01.932712 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:01Z","lastTransitionTime":"2026-02-22T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.005872 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 21:16:36.035813553 +0000 UTC Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.034998 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.035057 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.035069 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.035085 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.035096 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:02Z","lastTransitionTime":"2026-02-22T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.077502 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:02 crc kubenswrapper[4857]: E0222 00:07:02.077646 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.137175 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.137214 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.137226 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.137243 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.137254 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:02Z","lastTransitionTime":"2026-02-22T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.238926 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.238959 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.238972 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.238989 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.239001 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:02Z","lastTransitionTime":"2026-02-22T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.295067 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" event={"ID":"fec51cf7-f11e-4829-bcd2-08ac17cdce91","Type":"ContainerDied","Data":"e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d"} Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.295072 4857 generic.go:334] "Generic (PLEG): container finished" podID="fec51cf7-f11e-4829-bcd2-08ac17cdce91" containerID="e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d" exitCode=0 Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.306266 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerStarted","Data":"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364"} Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.312078 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:02Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.337522 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:02Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.341435 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.341498 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.341518 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.341558 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.341579 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:02Z","lastTransitionTime":"2026-02-22T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.351843 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:02Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.369989 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:02Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.389844 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:02Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.400787 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:02Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.413283 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:02Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.423811 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:02Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.434387 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:02Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.444760 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.444823 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.444834 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.444848 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.444860 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:02Z","lastTransitionTime":"2026-02-22T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.445328 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:02Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.463992 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:02Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.477371 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:02Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.490097 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:02Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.502773 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:02Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.518208 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:02Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.546484 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.546519 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.546528 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.546541 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.546551 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:02Z","lastTransitionTime":"2026-02-22T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.649206 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.649254 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.649262 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.649277 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.649290 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:02Z","lastTransitionTime":"2026-02-22T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.751788 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.751824 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.751832 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.751846 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.751859 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:02Z","lastTransitionTime":"2026-02-22T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.802352 4857 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.854935 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.855009 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.855025 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.855076 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.855095 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:02Z","lastTransitionTime":"2026-02-22T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.957463 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.957500 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.957511 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.957528 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:02 crc kubenswrapper[4857]: I0222 00:07:02.957540 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:02Z","lastTransitionTime":"2026-02-22T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.006111 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 16:51:35.159351078 +0000 UTC Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.060739 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.060783 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.060804 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.060829 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.060845 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:03Z","lastTransitionTime":"2026-02-22T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.079176 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:03 crc kubenswrapper[4857]: E0222 00:07:03.079350 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.080009 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:03 crc kubenswrapper[4857]: E0222 00:07:03.080149 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.093132 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.104286 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.118571 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.130761 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.146929 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.159656 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.163662 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.163687 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.163695 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.163709 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.163719 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:03Z","lastTransitionTime":"2026-02-22T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.171815 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.182167 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.194178 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.205799 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.231296 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.244937 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.261385 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.265369 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.265404 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.265412 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.265426 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.265435 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:03Z","lastTransitionTime":"2026-02-22T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.279823 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.290885 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.315165 4857 generic.go:334] "Generic (PLEG): container finished" podID="fec51cf7-f11e-4829-bcd2-08ac17cdce91" containerID="07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e" exitCode=0 Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.315245 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" event={"ID":"fec51cf7-f11e-4829-bcd2-08ac17cdce91","Type":"ContainerDied","Data":"07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e"} Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.334941 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.346150 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.361866 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.370975 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.371019 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.371077 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.371102 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.371121 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:03Z","lastTransitionTime":"2026-02-22T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.380124 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.393895 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.410388 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.426656 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.441184 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.452078 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.462791 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.473579 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.473613 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.473622 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.473931 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.474279 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:03Z","lastTransitionTime":"2026-02-22T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.475502 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.498353 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.512847 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.524703 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.542058 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.576992 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.577057 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.577070 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.577085 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.577095 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:03Z","lastTransitionTime":"2026-02-22T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.679457 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.679488 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.679496 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.679508 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.679518 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:03Z","lastTransitionTime":"2026-02-22T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.781697 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.781728 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.781736 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.781749 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.781758 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:03Z","lastTransitionTime":"2026-02-22T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.883681 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.883712 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.883720 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.883732 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.883743 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:03Z","lastTransitionTime":"2026-02-22T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.986213 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.986251 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.986261 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.986275 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:03 crc kubenswrapper[4857]: I0222 00:07:03.986283 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:03Z","lastTransitionTime":"2026-02-22T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.007108 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 09:56:15.879245853 +0000 UTC Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.076975 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:04 crc kubenswrapper[4857]: E0222 00:07:04.077160 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.088755 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.088810 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.088827 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.088849 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.088867 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:04Z","lastTransitionTime":"2026-02-22T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.191356 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.191402 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.191414 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.191431 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.191442 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:04Z","lastTransitionTime":"2026-02-22T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.293901 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.293958 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.293971 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.293988 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.294000 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:04Z","lastTransitionTime":"2026-02-22T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.322814 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerStarted","Data":"a5c28df794f46cd9f7654ad21498171cf5a48d71c2d47957907cfc2aa5422b35"} Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.323016 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.327079 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" event={"ID":"fec51cf7-f11e-4829-bcd2-08ac17cdce91","Type":"ContainerStarted","Data":"b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca"} Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.337902 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.358535 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5c28df794f46cd9f7654ad21498171cf5a48d71c2d47957907cfc2aa5422b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.359337 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.370197 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.386270 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.396536 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.396578 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.396589 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.396605 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.396618 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:04Z","lastTransitionTime":"2026-02-22T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.401430 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.412213 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.423371 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.436603 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.448349 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.469186 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.482745 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.493580 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.498471 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.498500 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.498511 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.498527 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.498538 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:04Z","lastTransitionTime":"2026-02-22T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.506369 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.516982 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.531185 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.544464 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.557150 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.566838 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.578862 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.589053 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.599228 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.600470 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.600491 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.600501 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.600513 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.600522 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:04Z","lastTransitionTime":"2026-02-22T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.611351 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.622603 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.637694 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.655797 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.670390 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.682461 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.695264 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.702753 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.702798 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.702807 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.702824 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.702837 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:04Z","lastTransitionTime":"2026-02-22T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.714016 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5c28df794f46cd9f7654ad21498171cf5a48d71c2d47957907cfc2aa5422b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.724344 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:04Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.804783 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.804825 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.804837 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.804851 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.804863 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:04Z","lastTransitionTime":"2026-02-22T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.907602 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.907647 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.907661 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.907681 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:04 crc kubenswrapper[4857]: I0222 00:07:04.907691 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:04Z","lastTransitionTime":"2026-02-22T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.008303 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 02:14:17.031628951 +0000 UTC Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.009842 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.009879 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.009888 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.009904 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.009915 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:05Z","lastTransitionTime":"2026-02-22T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.077495 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.077554 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:05 crc kubenswrapper[4857]: E0222 00:07:05.077624 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:05 crc kubenswrapper[4857]: E0222 00:07:05.077688 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.112230 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.112266 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.112276 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.112288 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.112297 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:05Z","lastTransitionTime":"2026-02-22T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.214492 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.214525 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.214534 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.214546 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.214555 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:05Z","lastTransitionTime":"2026-02-22T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.316421 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.316464 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.316472 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.316485 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.316493 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:05Z","lastTransitionTime":"2026-02-22T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.329413 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.329821 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.350105 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.369242 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.387388 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.401521 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.411535 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.418044 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.418081 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.418112 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.418128 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.418139 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:05Z","lastTransitionTime":"2026-02-22T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.422664 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.434945 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.454673 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.467530 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.478102 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.495533 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5c28df794f46cd9f7654ad21498171cf5a48d71c2d47957907cfc2aa5422b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.507583 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.517827 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.520244 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.520277 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.520288 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.520304 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.520316 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:05Z","lastTransitionTime":"2026-02-22T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.530373 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.544336 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.557755 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.623011 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.623060 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.623069 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.623083 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.623092 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:05Z","lastTransitionTime":"2026-02-22T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.732433 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.732506 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.732518 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.732533 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.732546 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:05Z","lastTransitionTime":"2026-02-22T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.836239 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.836288 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.836299 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.836314 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.836325 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:05Z","lastTransitionTime":"2026-02-22T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.937895 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.937931 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.937941 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.937966 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:05 crc kubenswrapper[4857]: I0222 00:07:05.937978 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:05Z","lastTransitionTime":"2026-02-22T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.008634 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 02:39:48.012886344 +0000 UTC Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.040384 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.040416 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.040424 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.040441 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.040453 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:06Z","lastTransitionTime":"2026-02-22T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.076976 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:06 crc kubenswrapper[4857]: E0222 00:07:06.077163 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.142626 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.142677 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.142689 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.142707 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.142750 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:06Z","lastTransitionTime":"2026-02-22T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.245553 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.245589 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.245598 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.245611 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.245620 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:06Z","lastTransitionTime":"2026-02-22T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.332775 4857 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.348547 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.348612 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.348625 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.348645 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.348682 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:06Z","lastTransitionTime":"2026-02-22T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.450861 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.450906 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.450914 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.450928 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.450938 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:06Z","lastTransitionTime":"2026-02-22T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.553615 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.553661 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.553673 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.553692 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.553709 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:06Z","lastTransitionTime":"2026-02-22T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.655917 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.655961 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.655975 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.655994 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.656016 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:06Z","lastTransitionTime":"2026-02-22T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.758399 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.758442 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.758452 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.758467 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.758478 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:06Z","lastTransitionTime":"2026-02-22T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.861104 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.861144 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.861153 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.861168 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.861178 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:06Z","lastTransitionTime":"2026-02-22T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.963958 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.964003 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.964016 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.964050 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:06 crc kubenswrapper[4857]: I0222 00:07:06.964064 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:06Z","lastTransitionTime":"2026-02-22T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.009120 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 21:57:22.2100384 +0000 UTC Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.068254 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.068315 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.068327 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.068345 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.068357 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:07Z","lastTransitionTime":"2026-02-22T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.077566 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:07 crc kubenswrapper[4857]: E0222 00:07:07.077777 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.078297 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:07 crc kubenswrapper[4857]: E0222 00:07:07.078484 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.171264 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.171348 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.171369 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.171897 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.171958 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:07Z","lastTransitionTime":"2026-02-22T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.275449 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.275497 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.275506 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.275521 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.275532 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:07Z","lastTransitionTime":"2026-02-22T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.340622 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovnkube-controller/0.log" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.344678 4857 generic.go:334] "Generic (PLEG): container finished" podID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerID="a5c28df794f46cd9f7654ad21498171cf5a48d71c2d47957907cfc2aa5422b35" exitCode=1 Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.344726 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerDied","Data":"a5c28df794f46cd9f7654ad21498171cf5a48d71c2d47957907cfc2aa5422b35"} Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.346396 4857 scope.go:117] "RemoveContainer" containerID="a5c28df794f46cd9f7654ad21498171cf5a48d71c2d47957907cfc2aa5422b35" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.364493 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:07Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.379145 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.379194 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.379204 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.379225 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.379235 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:07Z","lastTransitionTime":"2026-02-22T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.388360 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:07Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.407634 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:07Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.421379 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:07Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.436738 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:07Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.454860 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:07Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.477103 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:07Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.481723 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.481789 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.481813 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.481854 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.481883 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:07Z","lastTransitionTime":"2026-02-22T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.504017 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:07Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.527142 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:07Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.559218 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5c28df794f46cd9f7654ad21498171cf5a48d71c2d47957907cfc2aa5422b35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5c28df794f46cd9f7654ad21498171cf5a48d71c2d47957907cfc2aa5422b35\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:06Z\\\",\\\"message\\\":\\\"Node event handler 7\\\\nI0222 00:07:06.622562 6160 handler.go:208] Removed *v1.Node event handler 2\\\\nI0222 00:07:06.622630 6160 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0222 00:07:06.622685 6160 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0222 00:07:06.623120 6160 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0222 00:07:06.623160 6160 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0222 00:07:06.623204 6160 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0222 00:07:06.623256 6160 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0222 00:07:06.623288 6160 factory.go:656] Stopping watch factory\\\\nI0222 00:07:06.623317 6160 ovnkube.go:599] Stopped ovnkube\\\\nI0222 00:07:06.623353 6160 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0222 00:07:06.623379 6160 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0222 00:07:06.623402 6160 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0222 00:07:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:07Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.577500 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:07Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.583811 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.583836 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.583844 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.583858 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.583868 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:07Z","lastTransitionTime":"2026-02-22T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.591522 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:07Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.607864 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:07Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.628851 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:07Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.642341 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:07Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.686980 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.687029 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.687059 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.687078 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.687090 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:07Z","lastTransitionTime":"2026-02-22T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.790024 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.790092 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.790104 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.790124 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.790138 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:07Z","lastTransitionTime":"2026-02-22T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.892900 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.892956 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.892971 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.892996 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.893012 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:07Z","lastTransitionTime":"2026-02-22T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.995203 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.995244 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.995254 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.995267 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:07 crc kubenswrapper[4857]: I0222 00:07:07.995277 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:07Z","lastTransitionTime":"2026-02-22T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.009267 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 12:10:17.279997419 +0000 UTC Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.077277 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:08 crc kubenswrapper[4857]: E0222 00:07:08.077467 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.098414 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.098456 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.098470 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.098488 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.098505 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:08Z","lastTransitionTime":"2026-02-22T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.202698 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.202762 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.202781 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.202812 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.202834 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:08Z","lastTransitionTime":"2026-02-22T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.305742 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.305777 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.305785 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.305799 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.305809 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:08Z","lastTransitionTime":"2026-02-22T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.349633 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovnkube-controller/1.log" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.350141 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovnkube-controller/0.log" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.352670 4857 generic.go:334] "Generic (PLEG): container finished" podID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerID="81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd" exitCode=1 Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.352721 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerDied","Data":"81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd"} Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.352764 4857 scope.go:117] "RemoveContainer" containerID="a5c28df794f46cd9f7654ad21498171cf5a48d71c2d47957907cfc2aa5422b35" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.353423 4857 scope.go:117] "RemoveContainer" containerID="81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd" Feb 22 00:07:08 crc kubenswrapper[4857]: E0222 00:07:08.353697 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.376714 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.396533 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.409028 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.409089 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.409102 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.409120 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.409134 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:08Z","lastTransitionTime":"2026-02-22T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.411805 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.424585 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.447767 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.468279 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.489610 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.505817 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.511850 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.511906 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.511916 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.511936 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.511952 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:08Z","lastTransitionTime":"2026-02-22T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.521492 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.537972 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.561554 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.586633 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.609856 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.614614 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.614643 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.614656 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.614675 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.614688 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:08Z","lastTransitionTime":"2026-02-22T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.632952 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.657110 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5c28df794f46cd9f7654ad21498171cf5a48d71c2d47957907cfc2aa5422b35\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:06Z\\\",\\\"message\\\":\\\"Node event handler 7\\\\nI0222 00:07:06.622562 6160 handler.go:208] Removed *v1.Node event handler 2\\\\nI0222 00:07:06.622630 6160 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI0222 00:07:06.622685 6160 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0222 00:07:06.623120 6160 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0222 00:07:06.623160 6160 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0222 00:07:06.623204 6160 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0222 00:07:06.623256 6160 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0222 00:07:06.623288 6160 factory.go:656] Stopping watch factory\\\\nI0222 00:07:06.623317 6160 ovnkube.go:599] Stopped ovnkube\\\\nI0222 00:07:06.623353 6160 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0222 00:07:06.623379 6160 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0222 00:07:06.623402 6160 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0222 00:07:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:08Z\\\",\\\"message\\\":\\\"entity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240714 6287 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240760 6287 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI0222 00:07:08.240782 6287 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0222 00:07:08.240795 6287 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.267609 6287 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0222 00:07:08.267699 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0222 00:07:08.267808 6287 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0222 00:07:08.267864 6287 factory.go:656] Stopping watch factory\\\\nI0222 00:07:08.267895 6287 ovnkube.go:599] Stopped ovnkube\\\\nI0222 00:07:08.267935 6287 handler.go:208] Removed *v1.Node event handler 2\\\\nI0222 00:07:08.267952 6287 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0222 00:07:08.268117 6287 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.718572 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.718650 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.718665 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.718688 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.718702 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:08Z","lastTransitionTime":"2026-02-22T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.763623 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.763744 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.763843 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:08 crc kubenswrapper[4857]: E0222 00:07:08.763938 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:07:24.763888644 +0000 UTC m=+52.402617907 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:07:08 crc kubenswrapper[4857]: E0222 00:07:08.763989 4857 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 22 00:07:08 crc kubenswrapper[4857]: E0222 00:07:08.763942 4857 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 22 00:07:08 crc kubenswrapper[4857]: E0222 00:07:08.764104 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:24.764081209 +0000 UTC m=+52.402810492 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 22 00:07:08 crc kubenswrapper[4857]: E0222 00:07:08.764134 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:24.76412153 +0000 UTC m=+52.402850823 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.822021 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.822145 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.822172 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.822205 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.822229 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:08Z","lastTransitionTime":"2026-02-22T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.864869 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.864964 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:08 crc kubenswrapper[4857]: E0222 00:07:08.865174 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 22 00:07:08 crc kubenswrapper[4857]: E0222 00:07:08.865217 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 22 00:07:08 crc kubenswrapper[4857]: E0222 00:07:08.865237 4857 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:07:08 crc kubenswrapper[4857]: E0222 00:07:08.865317 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 22 00:07:08 crc kubenswrapper[4857]: E0222 00:07:08.865387 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 22 00:07:08 crc kubenswrapper[4857]: E0222 00:07:08.865406 4857 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:07:08 crc kubenswrapper[4857]: E0222 00:07:08.865336 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:24.865306889 +0000 UTC m=+52.504036172 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:07:08 crc kubenswrapper[4857]: E0222 00:07:08.865527 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:24.865494354 +0000 UTC m=+52.504223637 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.926011 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.926140 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.926159 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.926188 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:08 crc kubenswrapper[4857]: I0222 00:07:08.926212 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:08Z","lastTransitionTime":"2026-02-22T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.010137 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 02:57:01.48113739 +0000 UTC Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.030123 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.030204 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.030221 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.030256 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.030275 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:09Z","lastTransitionTime":"2026-02-22T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.066480 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.066559 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.066587 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.066624 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.066671 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:09Z","lastTransitionTime":"2026-02-22T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.076932 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.076935 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:09 crc kubenswrapper[4857]: E0222 00:07:09.077125 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:09 crc kubenswrapper[4857]: E0222 00:07:09.077346 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:09 crc kubenswrapper[4857]: E0222 00:07:09.094220 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.099080 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.099144 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.099160 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.099182 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.099199 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:09Z","lastTransitionTime":"2026-02-22T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:09 crc kubenswrapper[4857]: E0222 00:07:09.121168 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.126785 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.126858 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.126917 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.126952 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.126982 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:09Z","lastTransitionTime":"2026-02-22T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:09 crc kubenswrapper[4857]: E0222 00:07:09.152348 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.158700 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.158784 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.158815 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.158890 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.158919 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:09Z","lastTransitionTime":"2026-02-22T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:09 crc kubenswrapper[4857]: E0222 00:07:09.178338 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.183144 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.183226 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.183256 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.183295 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.183325 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:09Z","lastTransitionTime":"2026-02-22T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:09 crc kubenswrapper[4857]: E0222 00:07:09.208136 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: E0222 00:07:09.209077 4857 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.214474 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.214529 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.214546 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.214577 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.214594 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:09Z","lastTransitionTime":"2026-02-22T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.318180 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.318254 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.318272 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.318300 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.318324 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:09Z","lastTransitionTime":"2026-02-22T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.360770 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovnkube-controller/1.log" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.364200 4857 scope.go:117] "RemoveContainer" containerID="81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd" Feb 22 00:07:09 crc kubenswrapper[4857]: E0222 00:07:09.364353 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.386142 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.421598 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.421643 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.421653 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.421671 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.421462 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:08Z\\\",\\\"message\\\":\\\"entity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240714 6287 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240760 6287 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI0222 00:07:08.240782 6287 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0222 00:07:08.240795 6287 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.267609 6287 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0222 00:07:08.267699 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0222 00:07:08.267808 6287 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0222 00:07:08.267864 6287 factory.go:656] Stopping watch factory\\\\nI0222 00:07:08.267895 6287 ovnkube.go:599] Stopped ovnkube\\\\nI0222 00:07:08.267935 6287 handler.go:208] Removed *v1.Node event handler 2\\\\nI0222 00:07:08.267952 6287 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0222 00:07:08.268117 6287 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.421682 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:09Z","lastTransitionTime":"2026-02-22T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.443939 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.465901 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.483900 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.500151 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.519609 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.525357 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.525394 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.525408 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.525432 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.525447 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:09Z","lastTransitionTime":"2026-02-22T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.539820 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.560241 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.582988 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.621010 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.628527 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.628627 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.628653 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.628691 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.628750 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:09Z","lastTransitionTime":"2026-02-22T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.644345 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.667247 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.690903 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.711315 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:09Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.732838 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.732917 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.732938 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.732968 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.732990 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:09Z","lastTransitionTime":"2026-02-22T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.836158 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.836233 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.836252 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.836287 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.836313 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:09Z","lastTransitionTime":"2026-02-22T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.940211 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.940286 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.940306 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.940335 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:09 crc kubenswrapper[4857]: I0222 00:07:09.940358 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:09Z","lastTransitionTime":"2026-02-22T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.011091 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 23:53:48.049107923 +0000 UTC Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.044350 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.044425 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.044442 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.044480 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.044503 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:10Z","lastTransitionTime":"2026-02-22T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.077464 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:10 crc kubenswrapper[4857]: E0222 00:07:10.077665 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.148277 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.148356 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.148376 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.148408 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.148428 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:10Z","lastTransitionTime":"2026-02-22T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.252142 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.252219 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.252236 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.252265 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.252280 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:10Z","lastTransitionTime":"2026-02-22T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.356179 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.356290 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.356326 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.356367 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.356396 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:10Z","lastTransitionTime":"2026-02-22T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.461279 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.461348 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.461363 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.461390 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.461411 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:10Z","lastTransitionTime":"2026-02-22T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.567777 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.567866 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.567896 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.567936 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.567966 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:10Z","lastTransitionTime":"2026-02-22T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.621143 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh"] Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.621977 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.625364 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.625619 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.644564 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.669440 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:08Z\\\",\\\"message\\\":\\\"entity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240714 6287 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240760 6287 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI0222 00:07:08.240782 6287 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0222 00:07:08.240795 6287 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.267609 6287 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0222 00:07:08.267699 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0222 00:07:08.267808 6287 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0222 00:07:08.267864 6287 factory.go:656] Stopping watch factory\\\\nI0222 00:07:08.267895 6287 ovnkube.go:599] Stopped ovnkube\\\\nI0222 00:07:08.267935 6287 handler.go:208] Removed *v1.Node event handler 2\\\\nI0222 00:07:08.267952 6287 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0222 00:07:08.268117 6287 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.671539 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.671592 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.671610 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.671635 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.671652 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:10Z","lastTransitionTime":"2026-02-22T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.684325 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.692393 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8763e232-ea3f-473d-92df-a087589ad717-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cp7qh\" (UID: \"8763e232-ea3f-473d-92df-a087589ad717\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.692498 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxzd8\" (UniqueName: \"kubernetes.io/projected/8763e232-ea3f-473d-92df-a087589ad717-kube-api-access-rxzd8\") pod \"ovnkube-control-plane-749d76644c-cp7qh\" (UID: \"8763e232-ea3f-473d-92df-a087589ad717\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.692554 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8763e232-ea3f-473d-92df-a087589ad717-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cp7qh\" (UID: \"8763e232-ea3f-473d-92df-a087589ad717\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.692594 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8763e232-ea3f-473d-92df-a087589ad717-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cp7qh\" (UID: \"8763e232-ea3f-473d-92df-a087589ad717\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.698348 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.710543 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.729106 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.742165 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.752252 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.764908 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.774663 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.774710 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.774724 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.774743 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.774752 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:10Z","lastTransitionTime":"2026-02-22T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.784206 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.793971 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxzd8\" (UniqueName: \"kubernetes.io/projected/8763e232-ea3f-473d-92df-a087589ad717-kube-api-access-rxzd8\") pod \"ovnkube-control-plane-749d76644c-cp7qh\" (UID: \"8763e232-ea3f-473d-92df-a087589ad717\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.794028 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8763e232-ea3f-473d-92df-a087589ad717-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cp7qh\" (UID: \"8763e232-ea3f-473d-92df-a087589ad717\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.794106 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8763e232-ea3f-473d-92df-a087589ad717-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cp7qh\" (UID: \"8763e232-ea3f-473d-92df-a087589ad717\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.794164 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8763e232-ea3f-473d-92df-a087589ad717-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cp7qh\" (UID: \"8763e232-ea3f-473d-92df-a087589ad717\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.794764 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8763e232-ea3f-473d-92df-a087589ad717-env-overrides\") pod \"ovnkube-control-plane-749d76644c-cp7qh\" (UID: \"8763e232-ea3f-473d-92df-a087589ad717\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.795489 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8763e232-ea3f-473d-92df-a087589ad717-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-cp7qh\" (UID: \"8763e232-ea3f-473d-92df-a087589ad717\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.797727 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.804897 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8763e232-ea3f-473d-92df-a087589ad717-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-cp7qh\" (UID: \"8763e232-ea3f-473d-92df-a087589ad717\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.808561 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.818903 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.825348 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxzd8\" (UniqueName: \"kubernetes.io/projected/8763e232-ea3f-473d-92df-a087589ad717-kube-api-access-rxzd8\") pod \"ovnkube-control-plane-749d76644c-cp7qh\" (UID: \"8763e232-ea3f-473d-92df-a087589ad717\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.836336 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.861897 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.877598 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.877679 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.877699 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.877734 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.877753 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:10Z","lastTransitionTime":"2026-02-22T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.881853 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.942282 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.981431 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.981623 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.981796 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.981963 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:10 crc kubenswrapper[4857]: I0222 00:07:10.982186 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:10Z","lastTransitionTime":"2026-02-22T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.011402 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 13:02:40.818149978 +0000 UTC Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.076937 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:11 crc kubenswrapper[4857]: E0222 00:07:11.077094 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.077177 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:11 crc kubenswrapper[4857]: E0222 00:07:11.077484 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.084939 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.084977 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.084990 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.085006 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.085019 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:11Z","lastTransitionTime":"2026-02-22T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.189368 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.189440 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.189461 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.189493 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.189515 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:11Z","lastTransitionTime":"2026-02-22T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.293090 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.293145 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.293164 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.293190 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.293209 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:11Z","lastTransitionTime":"2026-02-22T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.375131 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" event={"ID":"8763e232-ea3f-473d-92df-a087589ad717","Type":"ContainerStarted","Data":"5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2"} Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.375216 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" event={"ID":"8763e232-ea3f-473d-92df-a087589ad717","Type":"ContainerStarted","Data":"a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98"} Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.375239 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" event={"ID":"8763e232-ea3f-473d-92df-a087589ad717","Type":"ContainerStarted","Data":"4bed5080a198bc040f8ac9f7df00a9964cba1b79880a2dcd2d956fcf39626403"} Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.388007 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-jnm7q"] Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.388925 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:11 crc kubenswrapper[4857]: E0222 00:07:11.389131 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.391613 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.396371 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.396471 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.396493 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.396532 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.396556 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:11Z","lastTransitionTime":"2026-02-22T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.419382 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.455983 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.481245 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.499928 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.499990 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.500007 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.500065 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.500093 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:11Z","lastTransitionTime":"2026-02-22T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.501836 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs\") pod \"network-metrics-daemon-jnm7q\" (UID: \"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\") " pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.501913 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74d6t\" (UniqueName: \"kubernetes.io/projected/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-kube-api-access-74d6t\") pod \"network-metrics-daemon-jnm7q\" (UID: \"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\") " pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.511219 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.531240 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.542402 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.602600 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74d6t\" (UniqueName: \"kubernetes.io/projected/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-kube-api-access-74d6t\") pod \"network-metrics-daemon-jnm7q\" (UID: \"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\") " pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.602704 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.602751 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.602767 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.602792 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.602813 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:11Z","lastTransitionTime":"2026-02-22T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:11 crc kubenswrapper[4857]: E0222 00:07:11.602840 4857 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 22 00:07:11 crc kubenswrapper[4857]: E0222 00:07:11.603152 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs podName:a0fb0fe8-ad7a-48a5-8a76-737cd5076f53 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:12.10310102 +0000 UTC m=+39.741830283 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs") pod "network-metrics-daemon-jnm7q" (UID: "a0fb0fe8-ad7a-48a5-8a76-737cd5076f53") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.602714 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs\") pod \"network-metrics-daemon-jnm7q\" (UID: \"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\") " pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.622019 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.636787 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74d6t\" (UniqueName: \"kubernetes.io/projected/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-kube-api-access-74d6t\") pod \"network-metrics-daemon-jnm7q\" (UID: \"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\") " pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.643837 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.683531 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:08Z\\\",\\\"message\\\":\\\"entity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240714 6287 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240760 6287 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI0222 00:07:08.240782 6287 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0222 00:07:08.240795 6287 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.267609 6287 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0222 00:07:08.267699 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0222 00:07:08.267808 6287 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0222 00:07:08.267864 6287 factory.go:656] Stopping watch factory\\\\nI0222 00:07:08.267895 6287 ovnkube.go:599] Stopped ovnkube\\\\nI0222 00:07:08.267935 6287 handler.go:208] Removed *v1.Node event handler 2\\\\nI0222 00:07:08.267952 6287 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0222 00:07:08.268117 6287 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.699930 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.705390 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.705446 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.705457 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.705479 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.705494 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:11Z","lastTransitionTime":"2026-02-22T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.716015 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.734291 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.749426 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.764155 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.776860 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.797338 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:08Z\\\",\\\"message\\\":\\\"entity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240714 6287 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240760 6287 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI0222 00:07:08.240782 6287 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0222 00:07:08.240795 6287 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.267609 6287 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0222 00:07:08.267699 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0222 00:07:08.267808 6287 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0222 00:07:08.267864 6287 factory.go:656] Stopping watch factory\\\\nI0222 00:07:08.267895 6287 ovnkube.go:599] Stopped ovnkube\\\\nI0222 00:07:08.267935 6287 handler.go:208] Removed *v1.Node event handler 2\\\\nI0222 00:07:08.267952 6287 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0222 00:07:08.268117 6287 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.807750 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.807802 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.807813 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.807830 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.807842 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:11Z","lastTransitionTime":"2026-02-22T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.815420 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.826346 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.837637 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.846330 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.857350 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.870674 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.885663 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.906309 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.910468 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.910844 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.910926 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.910990 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.911077 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:11Z","lastTransitionTime":"2026-02-22T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.917527 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.927731 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.937677 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.945489 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.955502 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.966228 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.976376 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:11 crc kubenswrapper[4857]: I0222 00:07:11.990139 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.012302 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 16:36:20.812478102 +0000 UTC Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.012686 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.012710 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.012719 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.012732 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.012741 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:12Z","lastTransitionTime":"2026-02-22T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.077504 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:12 crc kubenswrapper[4857]: E0222 00:07:12.077649 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.108342 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs\") pod \"network-metrics-daemon-jnm7q\" (UID: \"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\") " pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:12 crc kubenswrapper[4857]: E0222 00:07:12.108651 4857 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 22 00:07:12 crc kubenswrapper[4857]: E0222 00:07:12.108807 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs podName:a0fb0fe8-ad7a-48a5-8a76-737cd5076f53 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:13.108772633 +0000 UTC m=+40.747501896 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs") pod "network-metrics-daemon-jnm7q" (UID: "a0fb0fe8-ad7a-48a5-8a76-737cd5076f53") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.114879 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.114921 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.114932 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.114948 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.114959 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:12Z","lastTransitionTime":"2026-02-22T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.217840 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.217933 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.217954 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.217987 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.218009 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:12Z","lastTransitionTime":"2026-02-22T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.320729 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.320772 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.320780 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.320797 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.320808 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:12Z","lastTransitionTime":"2026-02-22T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.423697 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.423769 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.423818 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.423847 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.423867 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:12Z","lastTransitionTime":"2026-02-22T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.527127 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.527204 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.527226 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.527263 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.527289 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:12Z","lastTransitionTime":"2026-02-22T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.630612 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.630672 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.630682 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.630703 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.630715 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:12Z","lastTransitionTime":"2026-02-22T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.733828 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.733882 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.733898 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.733915 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.733927 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:12Z","lastTransitionTime":"2026-02-22T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.836731 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.836839 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.836866 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.836884 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.836894 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:12Z","lastTransitionTime":"2026-02-22T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.940708 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.940762 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.940774 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.940803 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:12 crc kubenswrapper[4857]: I0222 00:07:12.940819 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:12Z","lastTransitionTime":"2026-02-22T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.013440 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 22:12:47.716612653 +0000 UTC Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.044351 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.044407 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.044418 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.044437 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.044448 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:13Z","lastTransitionTime":"2026-02-22T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.077113 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.077176 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.077112 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:13 crc kubenswrapper[4857]: E0222 00:07:13.077409 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:13 crc kubenswrapper[4857]: E0222 00:07:13.077507 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:13 crc kubenswrapper[4857]: E0222 00:07:13.077686 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.091422 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.110209 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.122118 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs\") pod \"network-metrics-daemon-jnm7q\" (UID: \"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\") " pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:13 crc kubenswrapper[4857]: E0222 00:07:13.122278 4857 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 22 00:07:13 crc kubenswrapper[4857]: E0222 00:07:13.122374 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs podName:a0fb0fe8-ad7a-48a5-8a76-737cd5076f53 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:15.122351132 +0000 UTC m=+42.761080385 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs") pod "network-metrics-daemon-jnm7q" (UID: "a0fb0fe8-ad7a-48a5-8a76-737cd5076f53") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.131722 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.146174 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.146246 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.146266 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.146296 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.146319 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:13Z","lastTransitionTime":"2026-02-22T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.148372 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.165058 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.184335 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.206381 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.221390 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.242491 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.248324 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.248398 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.248422 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.248451 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.248469 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:13Z","lastTransitionTime":"2026-02-22T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.253847 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.267375 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.282110 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.294860 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.308622 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.325971 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:08Z\\\",\\\"message\\\":\\\"entity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240714 6287 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240760 6287 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI0222 00:07:08.240782 6287 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0222 00:07:08.240795 6287 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.267609 6287 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0222 00:07:08.267699 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0222 00:07:08.267808 6287 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0222 00:07:08.267864 6287 factory.go:656] Stopping watch factory\\\\nI0222 00:07:08.267895 6287 ovnkube.go:599] Stopped ovnkube\\\\nI0222 00:07:08.267935 6287 handler.go:208] Removed *v1.Node event handler 2\\\\nI0222 00:07:08.267952 6287 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0222 00:07:08.268117 6287 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.340179 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.351401 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.351462 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.351480 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.351508 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.351530 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:13Z","lastTransitionTime":"2026-02-22T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.352351 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.454429 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.454513 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.454534 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.454569 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.454590 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:13Z","lastTransitionTime":"2026-02-22T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.557156 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.557251 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.557275 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.557305 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.557326 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:13Z","lastTransitionTime":"2026-02-22T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.660918 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.661002 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.661022 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.661077 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.661100 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:13Z","lastTransitionTime":"2026-02-22T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.764243 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.764317 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.764336 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.764365 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.764384 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:13Z","lastTransitionTime":"2026-02-22T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.867515 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.867569 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.867587 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.867609 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.867624 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:13Z","lastTransitionTime":"2026-02-22T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.970880 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.970955 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.970982 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.971017 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:13 crc kubenswrapper[4857]: I0222 00:07:13.971083 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:13Z","lastTransitionTime":"2026-02-22T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.013866 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 21:17:34.718626736 +0000 UTC Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.075280 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.075394 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.075408 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.075430 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.075448 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:14Z","lastTransitionTime":"2026-02-22T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.076680 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:14 crc kubenswrapper[4857]: E0222 00:07:14.076896 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.177759 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.177797 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.177809 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.177825 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.177840 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:14Z","lastTransitionTime":"2026-02-22T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.280864 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.280966 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.280990 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.281016 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.281076 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:14Z","lastTransitionTime":"2026-02-22T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.385075 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.385129 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.385143 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.385161 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.385173 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:14Z","lastTransitionTime":"2026-02-22T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.488139 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.488169 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.488179 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.488191 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.488200 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:14Z","lastTransitionTime":"2026-02-22T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.591507 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.591555 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.591567 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.591588 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.591599 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:14Z","lastTransitionTime":"2026-02-22T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.695069 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.695166 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.695189 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.695220 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.695246 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:14Z","lastTransitionTime":"2026-02-22T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.799580 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.799649 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.799672 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.799700 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.799722 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:14Z","lastTransitionTime":"2026-02-22T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.902480 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.902572 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.902593 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.902619 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:14 crc kubenswrapper[4857]: I0222 00:07:14.902638 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:14Z","lastTransitionTime":"2026-02-22T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.006356 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.006504 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.006532 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.006622 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.006675 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:15Z","lastTransitionTime":"2026-02-22T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.014623 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 10:15:19.940371817 +0000 UTC Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.077563 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.077688 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:15 crc kubenswrapper[4857]: E0222 00:07:15.077790 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.077852 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:15 crc kubenswrapper[4857]: E0222 00:07:15.078220 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:15 crc kubenswrapper[4857]: E0222 00:07:15.078368 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.110272 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.110354 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.110378 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.110409 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.110438 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:15Z","lastTransitionTime":"2026-02-22T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.149492 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs\") pod \"network-metrics-daemon-jnm7q\" (UID: \"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\") " pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:15 crc kubenswrapper[4857]: E0222 00:07:15.149748 4857 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 22 00:07:15 crc kubenswrapper[4857]: E0222 00:07:15.150003 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs podName:a0fb0fe8-ad7a-48a5-8a76-737cd5076f53 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:19.149844711 +0000 UTC m=+46.788573994 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs") pod "network-metrics-daemon-jnm7q" (UID: "a0fb0fe8-ad7a-48a5-8a76-737cd5076f53") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.214286 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.214351 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.214375 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.214404 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.214424 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:15Z","lastTransitionTime":"2026-02-22T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.318985 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.319107 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.319133 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.319172 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.319196 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:15Z","lastTransitionTime":"2026-02-22T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.422882 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.423092 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.423121 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.423191 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.423214 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:15Z","lastTransitionTime":"2026-02-22T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.526894 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.526935 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.526944 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.526961 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.526974 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:15Z","lastTransitionTime":"2026-02-22T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.630304 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.630354 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.630364 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.630384 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.630397 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:15Z","lastTransitionTime":"2026-02-22T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.733253 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.733334 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.733347 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.733376 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.733402 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:15Z","lastTransitionTime":"2026-02-22T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.836897 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.836954 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.836969 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.836991 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.837006 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:15Z","lastTransitionTime":"2026-02-22T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.940630 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.940686 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.940699 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.940721 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:15 crc kubenswrapper[4857]: I0222 00:07:15.940735 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:15Z","lastTransitionTime":"2026-02-22T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.014943 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 01:23:19.695246068 +0000 UTC Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.047726 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.047791 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.047808 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.047837 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.047856 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:16Z","lastTransitionTime":"2026-02-22T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.076758 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:16 crc kubenswrapper[4857]: E0222 00:07:16.077078 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.151396 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.151450 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.151465 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.151486 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.151505 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:16Z","lastTransitionTime":"2026-02-22T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.254746 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.254786 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.254798 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.254817 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.254831 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:16Z","lastTransitionTime":"2026-02-22T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.357240 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.357282 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.357290 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.357304 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.357314 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:16Z","lastTransitionTime":"2026-02-22T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.402864 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.404865 4857 scope.go:117] "RemoveContainer" containerID="81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd" Feb 22 00:07:16 crc kubenswrapper[4857]: E0222 00:07:16.405225 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.460779 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.460853 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.460876 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.460906 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.460924 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:16Z","lastTransitionTime":"2026-02-22T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.564026 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.564093 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.564102 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.564116 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.564126 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:16Z","lastTransitionTime":"2026-02-22T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.668110 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.668180 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.668198 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.668228 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.668247 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:16Z","lastTransitionTime":"2026-02-22T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.770981 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.771017 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.771027 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.771056 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.771068 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:16Z","lastTransitionTime":"2026-02-22T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.874507 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.874583 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.874602 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.874629 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.874646 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:16Z","lastTransitionTime":"2026-02-22T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.977479 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.977754 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.977782 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.977833 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:16 crc kubenswrapper[4857]: I0222 00:07:16.977853 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:16Z","lastTransitionTime":"2026-02-22T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.015916 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 00:44:54.422955015 +0000 UTC Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.077702 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.077709 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.077774 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:17 crc kubenswrapper[4857]: E0222 00:07:17.078151 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:17 crc kubenswrapper[4857]: E0222 00:07:17.078342 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:17 crc kubenswrapper[4857]: E0222 00:07:17.078506 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.079348 4857 scope.go:117] "RemoveContainer" containerID="190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.084881 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.084925 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.084939 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.084956 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.084971 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:17Z","lastTransitionTime":"2026-02-22T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.188143 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.188199 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.188212 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.188234 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.188249 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:17Z","lastTransitionTime":"2026-02-22T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.290461 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.290512 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.290523 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.290537 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.290549 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:17Z","lastTransitionTime":"2026-02-22T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.393801 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.393882 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.393906 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.393939 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.393967 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:17Z","lastTransitionTime":"2026-02-22T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.400057 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.402788 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918"} Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.403471 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.426101 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:17Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.444029 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:17Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.460464 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:17Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.475079 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:17Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.495910 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:17Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.498584 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.498632 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.498652 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.498675 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.498688 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:17Z","lastTransitionTime":"2026-02-22T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.529442 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:17Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.548550 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:17Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.571288 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:17Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.587714 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:17Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.601554 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.601628 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.601651 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.601683 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.601702 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:17Z","lastTransitionTime":"2026-02-22T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.602403 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:17Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.626475 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:08Z\\\",\\\"message\\\":\\\"entity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240714 6287 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240760 6287 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI0222 00:07:08.240782 6287 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0222 00:07:08.240795 6287 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.267609 6287 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0222 00:07:08.267699 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0222 00:07:08.267808 6287 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0222 00:07:08.267864 6287 factory.go:656] Stopping watch factory\\\\nI0222 00:07:08.267895 6287 ovnkube.go:599] Stopped ovnkube\\\\nI0222 00:07:08.267935 6287 handler.go:208] Removed *v1.Node event handler 2\\\\nI0222 00:07:08.267952 6287 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0222 00:07:08.268117 6287 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:17Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.643024 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:17Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.656867 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:17Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.675107 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:17Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.690349 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:17Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.704551 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.704603 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.704614 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.704635 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.704648 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:17Z","lastTransitionTime":"2026-02-22T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.705166 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:17Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.720635 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:17Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.806754 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.806840 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.806855 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.806874 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.806886 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:17Z","lastTransitionTime":"2026-02-22T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.908765 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.908806 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.908818 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.908833 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:17 crc kubenswrapper[4857]: I0222 00:07:17.908843 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:17Z","lastTransitionTime":"2026-02-22T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.012196 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.012243 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.012252 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.012266 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.012278 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:18Z","lastTransitionTime":"2026-02-22T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.016567 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 05:12:43.667725516 +0000 UTC Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.077182 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:18 crc kubenswrapper[4857]: E0222 00:07:18.077336 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.114726 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.114785 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.114799 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.114819 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.114833 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:18Z","lastTransitionTime":"2026-02-22T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.218272 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.218321 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.218333 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.218350 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.218363 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:18Z","lastTransitionTime":"2026-02-22T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.321301 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.321357 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.321373 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.321396 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.321413 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:18Z","lastTransitionTime":"2026-02-22T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.424359 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.424402 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.424410 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.424424 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.424433 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:18Z","lastTransitionTime":"2026-02-22T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.528275 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.528345 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.528364 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.528391 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.528413 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:18Z","lastTransitionTime":"2026-02-22T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.631281 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.631319 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.631326 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.631342 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.631350 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:18Z","lastTransitionTime":"2026-02-22T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.733760 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.733797 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.733806 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.733818 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.733828 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:18Z","lastTransitionTime":"2026-02-22T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.836492 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.836531 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.836541 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.836555 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.836565 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:18Z","lastTransitionTime":"2026-02-22T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.939152 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.939199 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.939209 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.939224 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:18 crc kubenswrapper[4857]: I0222 00:07:18.939233 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:18Z","lastTransitionTime":"2026-02-22T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.017447 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 06:51:48.822741387 +0000 UTC Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.041596 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.041648 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.041661 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.041680 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.041694 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:19Z","lastTransitionTime":"2026-02-22T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.077487 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.077547 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.077605 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:19 crc kubenswrapper[4857]: E0222 00:07:19.077759 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:19 crc kubenswrapper[4857]: E0222 00:07:19.077848 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:19 crc kubenswrapper[4857]: E0222 00:07:19.078023 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.144168 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.144244 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.144258 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.144277 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.144293 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:19Z","lastTransitionTime":"2026-02-22T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.197855 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs\") pod \"network-metrics-daemon-jnm7q\" (UID: \"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\") " pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:19 crc kubenswrapper[4857]: E0222 00:07:19.197999 4857 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 22 00:07:19 crc kubenswrapper[4857]: E0222 00:07:19.198081 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs podName:a0fb0fe8-ad7a-48a5-8a76-737cd5076f53 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:27.198067075 +0000 UTC m=+54.836796328 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs") pod "network-metrics-daemon-jnm7q" (UID: "a0fb0fe8-ad7a-48a5-8a76-737cd5076f53") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.246838 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.246906 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.246919 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.246939 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.246952 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:19Z","lastTransitionTime":"2026-02-22T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.348578 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.348617 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.348625 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.348639 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.348648 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:19Z","lastTransitionTime":"2026-02-22T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.450644 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.450694 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.450706 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.450723 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.450737 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:19Z","lastTransitionTime":"2026-02-22T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.538830 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.538877 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.538888 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.538902 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.538911 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:19Z","lastTransitionTime":"2026-02-22T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:19 crc kubenswrapper[4857]: E0222 00:07:19.552157 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:19Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.557559 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.557606 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.557645 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.557666 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.557701 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:19Z","lastTransitionTime":"2026-02-22T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:19 crc kubenswrapper[4857]: E0222 00:07:19.572896 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:19Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.577810 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.577878 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.577898 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.577925 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.577983 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:19Z","lastTransitionTime":"2026-02-22T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:19 crc kubenswrapper[4857]: E0222 00:07:19.591611 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:19Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.595357 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.595435 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.595456 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.595478 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.595493 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:19Z","lastTransitionTime":"2026-02-22T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:19 crc kubenswrapper[4857]: E0222 00:07:19.608392 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:19Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.612631 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.612686 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.612703 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.612724 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.612742 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:19Z","lastTransitionTime":"2026-02-22T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:19 crc kubenswrapper[4857]: E0222 00:07:19.627945 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:19Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:19 crc kubenswrapper[4857]: E0222 00:07:19.628140 4857 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.630091 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.630135 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.630148 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.630165 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.630178 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:19Z","lastTransitionTime":"2026-02-22T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.732677 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.732727 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.732739 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.732756 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.732770 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:19Z","lastTransitionTime":"2026-02-22T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.835736 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.835819 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.835842 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.835871 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.835896 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:19Z","lastTransitionTime":"2026-02-22T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.939540 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.939619 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.939643 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.939675 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:19 crc kubenswrapper[4857]: I0222 00:07:19.939698 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:19Z","lastTransitionTime":"2026-02-22T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.017830 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 18:50:17.819625099 +0000 UTC Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.045118 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.045185 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.045198 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.045226 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.045242 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:20Z","lastTransitionTime":"2026-02-22T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.076716 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:20 crc kubenswrapper[4857]: E0222 00:07:20.076953 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.147323 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.147368 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.147380 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.147397 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.147409 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:20Z","lastTransitionTime":"2026-02-22T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.250529 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.250573 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.250589 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.250613 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.250630 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:20Z","lastTransitionTime":"2026-02-22T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.353368 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.353421 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.353433 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.353452 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.353465 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:20Z","lastTransitionTime":"2026-02-22T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.456160 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.456204 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.456215 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.456255 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.456269 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:20Z","lastTransitionTime":"2026-02-22T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.559072 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.559145 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.559168 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.559199 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.559224 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:20Z","lastTransitionTime":"2026-02-22T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.662247 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.662307 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.662323 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.662345 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.662362 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:20Z","lastTransitionTime":"2026-02-22T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.765364 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.765532 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.765554 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.765576 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.765594 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:20Z","lastTransitionTime":"2026-02-22T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.868600 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.868662 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.868679 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.868703 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.868719 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:20Z","lastTransitionTime":"2026-02-22T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.971202 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.971275 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.971298 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.971332 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:20 crc kubenswrapper[4857]: I0222 00:07:20.971360 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:20Z","lastTransitionTime":"2026-02-22T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.019026 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 14:56:00.436707752 +0000 UTC Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.073762 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.073822 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.073839 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.073867 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.073887 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:21Z","lastTransitionTime":"2026-02-22T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.077204 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.077528 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:21 crc kubenswrapper[4857]: E0222 00:07:21.077559 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.077600 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:21 crc kubenswrapper[4857]: E0222 00:07:21.077798 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:21 crc kubenswrapper[4857]: E0222 00:07:21.078225 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.176915 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.176986 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.177003 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.177028 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.177084 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:21Z","lastTransitionTime":"2026-02-22T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.280252 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.280380 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.280403 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.280433 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.280456 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:21Z","lastTransitionTime":"2026-02-22T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.383585 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.383672 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.383689 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.383714 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.383732 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:21Z","lastTransitionTime":"2026-02-22T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.487696 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.487992 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.488005 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.488022 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.488055 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:21Z","lastTransitionTime":"2026-02-22T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.590664 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.590707 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.590717 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.590733 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.590745 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:21Z","lastTransitionTime":"2026-02-22T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.693405 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.693438 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.693451 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.693472 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.693486 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:21Z","lastTransitionTime":"2026-02-22T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.796598 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.796636 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.796648 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.796666 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.796679 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:21Z","lastTransitionTime":"2026-02-22T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.899142 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.899194 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.899202 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.899216 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:21 crc kubenswrapper[4857]: I0222 00:07:21.899225 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:21Z","lastTransitionTime":"2026-02-22T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.001953 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.002015 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.002026 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.002068 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.002081 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:22Z","lastTransitionTime":"2026-02-22T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.019463 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 07:09:47.393059784 +0000 UTC Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.076690 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:22 crc kubenswrapper[4857]: E0222 00:07:22.076820 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.104720 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.105009 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.105065 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.105098 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.105117 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:22Z","lastTransitionTime":"2026-02-22T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.207627 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.207679 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.207690 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.207708 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.207719 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:22Z","lastTransitionTime":"2026-02-22T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.310359 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.310433 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.310459 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.310488 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.310509 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:22Z","lastTransitionTime":"2026-02-22T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.413304 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.413347 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.413358 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.413378 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.413391 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:22Z","lastTransitionTime":"2026-02-22T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.442783 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.452290 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.456304 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:22Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.476129 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:22Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.488221 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:22Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.501002 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:22Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.514286 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:22Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.515633 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.515678 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.515689 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.515708 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.515723 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:22Z","lastTransitionTime":"2026-02-22T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.524837 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:22Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.536337 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:22Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.548062 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:22Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.559087 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:22Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.568680 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:22Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.579806 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:22Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.593471 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:22Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.617931 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.618006 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.618029 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.618115 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.618142 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:22Z","lastTransitionTime":"2026-02-22T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.620542 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:22Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.639374 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:22Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.648733 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:22Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.658539 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:22Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.679407 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:08Z\\\",\\\"message\\\":\\\"entity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240714 6287 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240760 6287 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI0222 00:07:08.240782 6287 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0222 00:07:08.240795 6287 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.267609 6287 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0222 00:07:08.267699 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0222 00:07:08.267808 6287 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0222 00:07:08.267864 6287 factory.go:656] Stopping watch factory\\\\nI0222 00:07:08.267895 6287 ovnkube.go:599] Stopped ovnkube\\\\nI0222 00:07:08.267935 6287 handler.go:208] Removed *v1.Node event handler 2\\\\nI0222 00:07:08.267952 6287 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0222 00:07:08.268117 6287 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:22Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.719882 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.719923 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.719931 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.719949 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.719958 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:22Z","lastTransitionTime":"2026-02-22T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.821815 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.821858 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.821867 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.821881 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.821891 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:22Z","lastTransitionTime":"2026-02-22T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.924659 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.924704 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.924718 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.924736 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:22 crc kubenswrapper[4857]: I0222 00:07:22.924751 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:22Z","lastTransitionTime":"2026-02-22T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.019597 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 16:17:26.302740442 +0000 UTC Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.027947 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.028010 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.028063 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.028096 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.028118 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:23Z","lastTransitionTime":"2026-02-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.076972 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.077124 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:23 crc kubenswrapper[4857]: E0222 00:07:23.077199 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.077255 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:23 crc kubenswrapper[4857]: E0222 00:07:23.077422 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:23 crc kubenswrapper[4857]: E0222 00:07:23.077542 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.088940 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea7b4f51-11b4-4b6b-a691-519f425b19cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a6a9ee3b161774e7dffa093ffbff0c3fb4202d2ca7a0351ae11bd2d5de064c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f62ed2ecd8930d7f3e899f7891c87b94f1e117a23588b2ce65581a3bd37f225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9920d33f76b36664452e744820c9fb71956c5773ca81b750d6202ed6cf08c4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:23Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.109663 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:23Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.129527 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:23Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.130714 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.130780 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.130803 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.130827 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.130844 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:23Z","lastTransitionTime":"2026-02-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.146762 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:23Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.162387 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:23Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.177319 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:23Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.192400 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:23Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.216708 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:23Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.233832 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.233871 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.233883 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.233901 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.233914 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:23Z","lastTransitionTime":"2026-02-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.236312 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:23Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.248345 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:23Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.262919 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:23Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.288137 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:08Z\\\",\\\"message\\\":\\\"entity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240714 6287 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240760 6287 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI0222 00:07:08.240782 6287 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0222 00:07:08.240795 6287 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.267609 6287 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0222 00:07:08.267699 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0222 00:07:08.267808 6287 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0222 00:07:08.267864 6287 factory.go:656] Stopping watch factory\\\\nI0222 00:07:08.267895 6287 ovnkube.go:599] Stopped ovnkube\\\\nI0222 00:07:08.267935 6287 handler.go:208] Removed *v1.Node event handler 2\\\\nI0222 00:07:08.267952 6287 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0222 00:07:08.268117 6287 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:23Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.304122 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:23Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.316502 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:23Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.332025 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:23Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.336309 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.336368 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.336396 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.336429 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.336451 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:23Z","lastTransitionTime":"2026-02-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.343170 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:23Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.359898 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:23Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.372139 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:23Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.438820 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.438874 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.438893 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.438914 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.438933 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:23Z","lastTransitionTime":"2026-02-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.541077 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.541137 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.541163 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.541193 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.541217 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:23Z","lastTransitionTime":"2026-02-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.644207 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.644278 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.644302 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.644330 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.644352 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:23Z","lastTransitionTime":"2026-02-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.746630 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.746695 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.746713 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.746737 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.746754 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:23Z","lastTransitionTime":"2026-02-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.849468 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.849542 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.849565 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.849593 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.849611 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:23Z","lastTransitionTime":"2026-02-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.952171 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.952260 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.952286 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.952319 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:23 crc kubenswrapper[4857]: I0222 00:07:23.952357 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:23Z","lastTransitionTime":"2026-02-22T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.020568 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 01:06:46.474336379 +0000 UTC Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.055095 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.055141 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.055159 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.055185 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.055204 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:24Z","lastTransitionTime":"2026-02-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.076914 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:24 crc kubenswrapper[4857]: E0222 00:07:24.077120 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.158439 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.158509 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.158527 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.158585 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.158596 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:24Z","lastTransitionTime":"2026-02-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.261311 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.261359 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.261374 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.261392 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.261406 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:24Z","lastTransitionTime":"2026-02-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.364655 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.364701 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.364713 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.364730 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.364742 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:24Z","lastTransitionTime":"2026-02-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.467783 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.467830 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.467847 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.467869 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.467889 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:24Z","lastTransitionTime":"2026-02-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.571337 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.571399 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.571423 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.571452 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.571474 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:24Z","lastTransitionTime":"2026-02-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.674516 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.674574 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.674590 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.674613 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.674629 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:24Z","lastTransitionTime":"2026-02-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.777667 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.777734 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.777760 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.777791 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.777815 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:24Z","lastTransitionTime":"2026-02-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.856025 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.856258 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:24 crc kubenswrapper[4857]: E0222 00:07:24.856326 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:07:56.856288931 +0000 UTC m=+84.495018234 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.856379 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:24 crc kubenswrapper[4857]: E0222 00:07:24.856441 4857 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 22 00:07:24 crc kubenswrapper[4857]: E0222 00:07:24.856580 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:56.856556129 +0000 UTC m=+84.495285432 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 22 00:07:24 crc kubenswrapper[4857]: E0222 00:07:24.856600 4857 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 22 00:07:24 crc kubenswrapper[4857]: E0222 00:07:24.856723 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:56.856701633 +0000 UTC m=+84.495430926 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.880767 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.880805 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.880816 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.880831 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.880842 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:24Z","lastTransitionTime":"2026-02-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.957660 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.957732 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:24 crc kubenswrapper[4857]: E0222 00:07:24.957868 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 22 00:07:24 crc kubenswrapper[4857]: E0222 00:07:24.957889 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 22 00:07:24 crc kubenswrapper[4857]: E0222 00:07:24.957901 4857 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:07:24 crc kubenswrapper[4857]: E0222 00:07:24.957943 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 22 00:07:24 crc kubenswrapper[4857]: E0222 00:07:24.957987 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 22 00:07:24 crc kubenswrapper[4857]: E0222 00:07:24.958010 4857 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:07:24 crc kubenswrapper[4857]: E0222 00:07:24.957955 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:56.957939873 +0000 UTC m=+84.596669126 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:07:24 crc kubenswrapper[4857]: E0222 00:07:24.958154 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:56.958118988 +0000 UTC m=+84.596848281 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.982973 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.983028 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.983061 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.983077 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:24 crc kubenswrapper[4857]: I0222 00:07:24.983089 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:24Z","lastTransitionTime":"2026-02-22T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.021663 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 11:43:34.59253686 +0000 UTC Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.076516 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.076517 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:25 crc kubenswrapper[4857]: E0222 00:07:25.076705 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.076739 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:25 crc kubenswrapper[4857]: E0222 00:07:25.076844 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:25 crc kubenswrapper[4857]: E0222 00:07:25.077000 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.086485 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.086525 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.086535 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.086550 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.086563 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:25Z","lastTransitionTime":"2026-02-22T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.188929 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.188962 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.188971 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.188984 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.188993 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:25Z","lastTransitionTime":"2026-02-22T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.291887 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.291941 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.291959 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.291981 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.291999 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:25Z","lastTransitionTime":"2026-02-22T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.395187 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.395297 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.395320 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.395344 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.395362 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:25Z","lastTransitionTime":"2026-02-22T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.498262 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.498317 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.498334 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.498359 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.498376 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:25Z","lastTransitionTime":"2026-02-22T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.601717 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.601796 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.601821 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.601856 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.601881 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:25Z","lastTransitionTime":"2026-02-22T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.706222 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.706297 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.706323 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.706355 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.706382 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:25Z","lastTransitionTime":"2026-02-22T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.809351 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.809417 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.809435 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.809459 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.809478 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:25Z","lastTransitionTime":"2026-02-22T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.912355 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.912497 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.912561 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.912585 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:25 crc kubenswrapper[4857]: I0222 00:07:25.912601 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:25Z","lastTransitionTime":"2026-02-22T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.015679 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.015737 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.015753 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.015775 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.015791 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:26Z","lastTransitionTime":"2026-02-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.022247 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 12:02:16.127583064 +0000 UTC Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.076943 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:26 crc kubenswrapper[4857]: E0222 00:07:26.077212 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.119546 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.119609 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.119629 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.119656 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.119676 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:26Z","lastTransitionTime":"2026-02-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.222637 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.222701 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.222718 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.222743 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.222759 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:26Z","lastTransitionTime":"2026-02-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.326455 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.326545 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.326568 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.326599 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.326620 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:26Z","lastTransitionTime":"2026-02-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.429368 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.429524 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.429555 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.429585 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.429606 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:26Z","lastTransitionTime":"2026-02-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.533280 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.533340 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.533361 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.533386 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.533406 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:26Z","lastTransitionTime":"2026-02-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.635936 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.635999 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.636030 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.636073 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.636086 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:26Z","lastTransitionTime":"2026-02-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.740396 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.740473 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.740491 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.740521 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.740542 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:26Z","lastTransitionTime":"2026-02-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.843465 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.843536 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.843558 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.843589 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.843614 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:26Z","lastTransitionTime":"2026-02-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.946651 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.946735 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.946758 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.946786 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:26 crc kubenswrapper[4857]: I0222 00:07:26.946806 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:26Z","lastTransitionTime":"2026-02-22T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.023291 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 05:03:08.261024822 +0000 UTC Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.049108 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.049143 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.049152 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.049167 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.049176 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:27Z","lastTransitionTime":"2026-02-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.077449 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.077449 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:27 crc kubenswrapper[4857]: E0222 00:07:27.077588 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:27 crc kubenswrapper[4857]: E0222 00:07:27.077793 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.077456 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:27 crc kubenswrapper[4857]: E0222 00:07:27.077882 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.150930 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.150992 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.151009 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.151067 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.151091 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:27Z","lastTransitionTime":"2026-02-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.253409 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.253475 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.253497 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.253527 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.253548 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:27Z","lastTransitionTime":"2026-02-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.285088 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs\") pod \"network-metrics-daemon-jnm7q\" (UID: \"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\") " pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:27 crc kubenswrapper[4857]: E0222 00:07:27.285224 4857 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 22 00:07:27 crc kubenswrapper[4857]: E0222 00:07:27.285285 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs podName:a0fb0fe8-ad7a-48a5-8a76-737cd5076f53 nodeName:}" failed. No retries permitted until 2026-02-22 00:07:43.285269089 +0000 UTC m=+70.923998342 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs") pod "network-metrics-daemon-jnm7q" (UID: "a0fb0fe8-ad7a-48a5-8a76-737cd5076f53") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.355540 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.355604 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.355618 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.355641 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.355658 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:27Z","lastTransitionTime":"2026-02-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.457973 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.458013 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.458021 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.458048 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.458064 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:27Z","lastTransitionTime":"2026-02-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.560179 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.560221 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.560232 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.560248 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.560260 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:27Z","lastTransitionTime":"2026-02-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.663181 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.663282 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.663307 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.663379 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.663415 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:27Z","lastTransitionTime":"2026-02-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.768276 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.769105 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.769229 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.769319 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.769373 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:27Z","lastTransitionTime":"2026-02-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.872234 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.872269 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.872280 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.872300 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.872313 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:27Z","lastTransitionTime":"2026-02-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.975418 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.975452 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.975461 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.975477 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:27 crc kubenswrapper[4857]: I0222 00:07:27.975486 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:27Z","lastTransitionTime":"2026-02-22T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.023423 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 21:32:25.596919145 +0000 UTC Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.077533 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:28 crc kubenswrapper[4857]: E0222 00:07:28.077673 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.078436 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.078466 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.078476 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.078489 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.078498 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:28Z","lastTransitionTime":"2026-02-22T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.079141 4857 scope.go:117] "RemoveContainer" containerID="81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.180810 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.181135 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.181143 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.181157 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.181183 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:28Z","lastTransitionTime":"2026-02-22T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.284243 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.284301 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.284312 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.284329 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.284364 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:28Z","lastTransitionTime":"2026-02-22T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.387318 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.387356 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.387366 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.387388 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.387399 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:28Z","lastTransitionTime":"2026-02-22T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.442446 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovnkube-controller/1.log" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.444791 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerStarted","Data":"13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2"} Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.445677 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.455909 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.468706 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.479424 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.488237 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.489441 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.489511 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.489523 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.489563 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.489578 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:28Z","lastTransitionTime":"2026-02-22T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.501456 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.512622 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.523384 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.541820 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.553278 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea7b4f51-11b4-4b6b-a691-519f425b19cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a6a9ee3b161774e7dffa093ffbff0c3fb4202d2ca7a0351ae11bd2d5de064c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f62ed2ecd8930d7f3e899f7891c87b94f1e117a23588b2ce65581a3bd37f225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9920d33f76b36664452e744820c9fb71956c5773ca81b750d6202ed6cf08c4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.567002 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.578416 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.589964 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.591155 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.591188 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.591200 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.591214 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.591224 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:28Z","lastTransitionTime":"2026-02-22T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.599790 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.609867 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.623198 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.632438 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.642228 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.660673 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:08Z\\\",\\\"message\\\":\\\"entity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240714 6287 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240760 6287 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI0222 00:07:08.240782 6287 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0222 00:07:08.240795 6287 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.267609 6287 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0222 00:07:08.267699 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0222 00:07:08.267808 6287 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0222 00:07:08.267864 6287 factory.go:656] Stopping watch factory\\\\nI0222 00:07:08.267895 6287 ovnkube.go:599] Stopped ovnkube\\\\nI0222 00:07:08.267935 6287 handler.go:208] Removed *v1.Node event handler 2\\\\nI0222 00:07:08.267952 6287 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0222 00:07:08.268117 6287 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.692940 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.692983 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.692994 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.693008 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.693017 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:28Z","lastTransitionTime":"2026-02-22T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.796494 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.796535 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.796546 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.796563 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.796574 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:28Z","lastTransitionTime":"2026-02-22T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.898735 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.898777 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.898786 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.898801 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:28 crc kubenswrapper[4857]: I0222 00:07:28.898810 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:28Z","lastTransitionTime":"2026-02-22T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.000530 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.000585 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.000602 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.000623 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.000642 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:29Z","lastTransitionTime":"2026-02-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.023537 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 05:22:34.692233412 +0000 UTC Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.077427 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.077526 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:29 crc kubenswrapper[4857]: E0222 00:07:29.077568 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:29 crc kubenswrapper[4857]: E0222 00:07:29.077655 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.077726 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:29 crc kubenswrapper[4857]: E0222 00:07:29.077780 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.102751 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.102789 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.102798 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.102811 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.102820 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:29Z","lastTransitionTime":"2026-02-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.206148 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.206188 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.206199 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.206215 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.206266 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:29Z","lastTransitionTime":"2026-02-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.308109 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.308170 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.308182 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.308197 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.308231 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:29Z","lastTransitionTime":"2026-02-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.410486 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.410523 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.410539 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.410554 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.410565 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:29Z","lastTransitionTime":"2026-02-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.448652 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovnkube-controller/2.log" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.449236 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovnkube-controller/1.log" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.451256 4857 generic.go:334] "Generic (PLEG): container finished" podID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerID="13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2" exitCode=1 Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.451294 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerDied","Data":"13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2"} Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.451328 4857 scope.go:117] "RemoveContainer" containerID="81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.452154 4857 scope.go:117] "RemoveContainer" containerID="13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2" Feb 22 00:07:29 crc kubenswrapper[4857]: E0222 00:07:29.452334 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.465698 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.475720 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.486650 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.496793 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.508195 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.512137 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.512169 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.512180 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.512195 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.512207 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:29Z","lastTransitionTime":"2026-02-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.524986 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.534810 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea7b4f51-11b4-4b6b-a691-519f425b19cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a6a9ee3b161774e7dffa093ffbff0c3fb4202d2ca7a0351ae11bd2d5de064c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f62ed2ecd8930d7f3e899f7891c87b94f1e117a23588b2ce65581a3bd37f225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9920d33f76b36664452e744820c9fb71956c5773ca81b750d6202ed6cf08c4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.545336 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.555524 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.565709 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.573449 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.581870 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.592846 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.602441 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.614837 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.614885 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.614903 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.614926 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.614940 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:29Z","lastTransitionTime":"2026-02-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.616719 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.637484 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://81bef6b7b90f934c32df059c2ec9546aba0a07492863a02b9a2b2988a53fa2bd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:08Z\\\",\\\"message\\\":\\\"entity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240714 6287 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.240760 6287 ovn.go:134] Ensuring zone local for Pod openshift-network-node-identity/network-node-identity-vrzqb in node crc\\\\nI0222 00:07:08.240782 6287 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI0222 00:07:08.240795 6287 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI0222 00:07:08.267609 6287 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0222 00:07:08.267699 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0222 00:07:08.267808 6287 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0222 00:07:08.267864 6287 factory.go:656] Stopping watch factory\\\\nI0222 00:07:08.267895 6287 ovnkube.go:599] Stopped ovnkube\\\\nI0222 00:07:08.267935 6287 handler.go:208] Removed *v1.Node event handler 2\\\\nI0222 00:07:08.267952 6287 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0222 00:07:08.268117 6287 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:28Z\\\",\\\"message\\\":\\\"ft-dns/node-resolver-2vqwv openshift-multus/multus-4lmlt openshift-multus/multus-additional-cni-plugins-zd2xg openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-image-registry/node-ca-lz8x8 openshift-kube-controller-manager/kube-controller-manager-crc]\\\\nF0222 00:07:28.806948 6573 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z]\\\\nI0222 00:07:28.806951 6573 services_controller.go:451] Built service openshift-kube-scheduler-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler-operator/metrics_TC\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.648172 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.659317 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.688673 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.688706 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.688716 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.688730 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.688740 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:29Z","lastTransitionTime":"2026-02-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:29 crc kubenswrapper[4857]: E0222 00:07:29.700565 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.703433 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.703465 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.703475 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.703494 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.703505 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:29Z","lastTransitionTime":"2026-02-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:29 crc kubenswrapper[4857]: E0222 00:07:29.714927 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.718000 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.718050 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.718061 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.718074 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.718083 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:29Z","lastTransitionTime":"2026-02-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:29 crc kubenswrapper[4857]: E0222 00:07:29.731118 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.734337 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.734387 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.734402 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.734423 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.734436 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:29Z","lastTransitionTime":"2026-02-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:29 crc kubenswrapper[4857]: E0222 00:07:29.744793 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.747990 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.748039 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.748049 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.748086 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.748106 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:29Z","lastTransitionTime":"2026-02-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:29 crc kubenswrapper[4857]: E0222 00:07:29.761421 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:29Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:29 crc kubenswrapper[4857]: E0222 00:07:29.761533 4857 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.762827 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.762845 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.762853 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.762866 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.762874 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:29Z","lastTransitionTime":"2026-02-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.864357 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.864382 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.864389 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.864402 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.864411 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:29Z","lastTransitionTime":"2026-02-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.966564 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.966646 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.966669 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.966698 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:29 crc kubenswrapper[4857]: I0222 00:07:29.966720 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:29Z","lastTransitionTime":"2026-02-22T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.024000 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-06 19:16:56.752926357 +0000 UTC Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.069746 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.069780 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.069788 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.069833 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.069845 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:30Z","lastTransitionTime":"2026-02-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.077458 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:30 crc kubenswrapper[4857]: E0222 00:07:30.077633 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.172205 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.172241 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.172252 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.172270 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.172281 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:30Z","lastTransitionTime":"2026-02-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.274586 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.274627 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.274643 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.274658 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.274668 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:30Z","lastTransitionTime":"2026-02-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.377076 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.377143 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.377157 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.377173 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.377186 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:30Z","lastTransitionTime":"2026-02-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.458030 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovnkube-controller/2.log" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.463186 4857 scope.go:117] "RemoveContainer" containerID="13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2" Feb 22 00:07:30 crc kubenswrapper[4857]: E0222 00:07:30.463455 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.477927 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:30Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.479336 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.479410 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.479433 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.479462 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.479484 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:30Z","lastTransitionTime":"2026-02-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.488665 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:30Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.503299 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:30Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.522542 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:30Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.533106 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:30Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.549398 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:30Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.563370 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:30Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.573804 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:30Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.582926 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.582959 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.582968 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.582984 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.582996 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:30Z","lastTransitionTime":"2026-02-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.584584 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:30Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.595676 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:30Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.605904 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:30Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.627372 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:30Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.641388 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea7b4f51-11b4-4b6b-a691-519f425b19cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a6a9ee3b161774e7dffa093ffbff0c3fb4202d2ca7a0351ae11bd2d5de064c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f62ed2ecd8930d7f3e899f7891c87b94f1e117a23588b2ce65581a3bd37f225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9920d33f76b36664452e744820c9fb71956c5773ca81b750d6202ed6cf08c4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:30Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.655743 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:30Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.668924 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:30Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.684855 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:30Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.685196 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.685231 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.685239 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.685253 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.685263 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:30Z","lastTransitionTime":"2026-02-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.711612 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:28Z\\\",\\\"message\\\":\\\"ft-dns/node-resolver-2vqwv openshift-multus/multus-4lmlt openshift-multus/multus-additional-cni-plugins-zd2xg openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-image-registry/node-ca-lz8x8 openshift-kube-controller-manager/kube-controller-manager-crc]\\\\nF0222 00:07:28.806948 6573 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z]\\\\nI0222 00:07:28.806951 6573 services_controller.go:451] Built service openshift-kube-scheduler-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler-operator/metrics_TC\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:30Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.725981 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:30Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.787573 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.787621 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.787635 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.787652 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.787664 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:30Z","lastTransitionTime":"2026-02-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.890037 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.890146 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.890159 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.890178 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.890190 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:30Z","lastTransitionTime":"2026-02-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.993621 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.993699 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.993724 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.993755 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:30 crc kubenswrapper[4857]: I0222 00:07:30.993781 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:30Z","lastTransitionTime":"2026-02-22T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.024293 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 03:32:16.337858035 +0000 UTC Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.076744 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.076761 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:31 crc kubenswrapper[4857]: E0222 00:07:31.076982 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.076775 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:31 crc kubenswrapper[4857]: E0222 00:07:31.076883 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:31 crc kubenswrapper[4857]: E0222 00:07:31.077275 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.096620 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.096668 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.096676 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.096688 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.096698 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:31Z","lastTransitionTime":"2026-02-22T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.163114 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.180582 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:31Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.198841 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.198888 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.198899 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.198917 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.198930 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:31Z","lastTransitionTime":"2026-02-22T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.210151 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:28Z\\\",\\\"message\\\":\\\"ft-dns/node-resolver-2vqwv openshift-multus/multus-4lmlt openshift-multus/multus-additional-cni-plugins-zd2xg openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-image-registry/node-ca-lz8x8 openshift-kube-controller-manager/kube-controller-manager-crc]\\\\nF0222 00:07:28.806948 6573 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z]\\\\nI0222 00:07:28.806951 6573 services_controller.go:451] Built service openshift-kube-scheduler-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler-operator/metrics_TC\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:31Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.222575 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:31Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.237142 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:31Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.252744 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:31Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.264978 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:31Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.281477 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:31Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.300225 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:31Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.302387 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.302443 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.302467 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.302500 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.302526 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:31Z","lastTransitionTime":"2026-02-22T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.315149 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea7b4f51-11b4-4b6b-a691-519f425b19cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a6a9ee3b161774e7dffa093ffbff0c3fb4202d2ca7a0351ae11bd2d5de064c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f62ed2ecd8930d7f3e899f7891c87b94f1e117a23588b2ce65581a3bd37f225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9920d33f76b36664452e744820c9fb71956c5773ca81b750d6202ed6cf08c4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:31Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.330362 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:31Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.348084 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:31Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.365228 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:31Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.378348 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:31Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.392449 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:31Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.405648 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.405710 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.405728 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.405750 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.405766 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:31Z","lastTransitionTime":"2026-02-22T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.408253 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:31Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.432367 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:31Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.449919 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:31Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.466921 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:31Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.508479 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.508749 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.509166 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.509522 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.509779 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:31Z","lastTransitionTime":"2026-02-22T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.613383 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.613448 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.613476 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.613508 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.613530 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:31Z","lastTransitionTime":"2026-02-22T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.717274 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.717385 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.717413 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.717450 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.717475 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:31Z","lastTransitionTime":"2026-02-22T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.819670 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.819736 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.819755 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.819779 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.819799 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:31Z","lastTransitionTime":"2026-02-22T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.922783 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.922945 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.922970 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.922996 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:31 crc kubenswrapper[4857]: I0222 00:07:31.923014 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:31Z","lastTransitionTime":"2026-02-22T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.024436 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 03:12:54.314164149 +0000 UTC Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.025828 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.025875 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.025888 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.025913 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.025938 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:32Z","lastTransitionTime":"2026-02-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.077129 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:32 crc kubenswrapper[4857]: E0222 00:07:32.077303 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.127987 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.128093 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.128109 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.128125 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.128134 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:32Z","lastTransitionTime":"2026-02-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.230003 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.230048 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.230074 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.230089 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.230100 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:32Z","lastTransitionTime":"2026-02-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.333231 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.333303 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.333322 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.333346 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.333363 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:32Z","lastTransitionTime":"2026-02-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.436363 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.436401 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.436422 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.436441 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.436453 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:32Z","lastTransitionTime":"2026-02-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.539654 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.539775 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.539800 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.539829 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.539852 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:32Z","lastTransitionTime":"2026-02-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.643519 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.643585 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.643603 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.643629 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.643647 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:32Z","lastTransitionTime":"2026-02-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.746532 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.746587 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.746622 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.746644 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.746656 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:32Z","lastTransitionTime":"2026-02-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.849553 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.849604 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.849618 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.849639 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.849654 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:32Z","lastTransitionTime":"2026-02-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.952342 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.952396 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.952405 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.952418 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:32 crc kubenswrapper[4857]: I0222 00:07:32.952432 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:32Z","lastTransitionTime":"2026-02-22T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.024928 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 03:26:05.349421947 +0000 UTC Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.054686 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.054733 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.054747 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.054768 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.054782 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:33Z","lastTransitionTime":"2026-02-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.077163 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.077226 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:33 crc kubenswrapper[4857]: E0222 00:07:33.077288 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.077156 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:33 crc kubenswrapper[4857]: E0222 00:07:33.077426 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:33 crc kubenswrapper[4857]: E0222 00:07:33.077583 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.090050 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea7b4f51-11b4-4b6b-a691-519f425b19cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a6a9ee3b161774e7dffa093ffbff0c3fb4202d2ca7a0351ae11bd2d5de064c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f62ed2ecd8930d7f3e899f7891c87b94f1e117a23588b2ce65581a3bd37f225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9920d33f76b36664452e744820c9fb71956c5773ca81b750d6202ed6cf08c4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:33Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.103874 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:33Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.115556 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:33Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.129910 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:33Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.141844 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:33Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.155246 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:33Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.157405 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.157482 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.157494 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.157512 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.157524 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:33Z","lastTransitionTime":"2026-02-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.168231 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:33Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.187169 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:33Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.200767 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:33Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.211388 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:33Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.224066 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:33Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.245086 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:28Z\\\",\\\"message\\\":\\\"ft-dns/node-resolver-2vqwv openshift-multus/multus-4lmlt openshift-multus/multus-additional-cni-plugins-zd2xg openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-image-registry/node-ca-lz8x8 openshift-kube-controller-manager/kube-controller-manager-crc]\\\\nF0222 00:07:28.806948 6573 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z]\\\\nI0222 00:07:28.806951 6573 services_controller.go:451] Built service openshift-kube-scheduler-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler-operator/metrics_TC\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:33Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.255360 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:33Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.259582 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.259627 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.259641 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.259658 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.259670 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:33Z","lastTransitionTime":"2026-02-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.264837 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:33Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.278274 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:33Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.288509 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:33Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.302191 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:33Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.319749 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:33Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.362057 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.362324 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.362430 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.362516 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.362618 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:33Z","lastTransitionTime":"2026-02-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.465674 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.465706 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.465716 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.465729 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.465738 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:33Z","lastTransitionTime":"2026-02-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.567408 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.567822 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.567844 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.567869 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.567885 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:33Z","lastTransitionTime":"2026-02-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.671025 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.671157 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.671179 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.671195 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.671205 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:33Z","lastTransitionTime":"2026-02-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.774136 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.774208 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.774228 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.774255 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.774273 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:33Z","lastTransitionTime":"2026-02-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.877006 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.877067 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.877077 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.877091 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.877102 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:33Z","lastTransitionTime":"2026-02-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.979546 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.979606 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.979624 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.979646 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:33 crc kubenswrapper[4857]: I0222 00:07:33.979663 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:33Z","lastTransitionTime":"2026-02-22T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.025548 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 18:51:16.181140004 +0000 UTC Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.077477 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:34 crc kubenswrapper[4857]: E0222 00:07:34.077659 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.082606 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.082680 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.082707 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.082738 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.082765 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:34Z","lastTransitionTime":"2026-02-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.185963 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.186009 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.186021 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.186044 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.186073 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:34Z","lastTransitionTime":"2026-02-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.289380 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.289451 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.289471 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.289495 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.289515 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:34Z","lastTransitionTime":"2026-02-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.392508 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.392543 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.392551 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.392564 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.392574 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:34Z","lastTransitionTime":"2026-02-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.495942 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.495986 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.495997 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.496012 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.496024 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:34Z","lastTransitionTime":"2026-02-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.599815 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.599889 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.599911 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.599936 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.599954 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:34Z","lastTransitionTime":"2026-02-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.703349 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.703409 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.703426 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.703450 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.703467 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:34Z","lastTransitionTime":"2026-02-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.806884 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.806933 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.806949 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.806971 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.806986 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:34Z","lastTransitionTime":"2026-02-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.909201 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.909253 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.909267 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.909304 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:34 crc kubenswrapper[4857]: I0222 00:07:34.909318 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:34Z","lastTransitionTime":"2026-02-22T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.011847 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.011902 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.011919 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.011944 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.011962 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:35Z","lastTransitionTime":"2026-02-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.026485 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 06:46:10.956136669 +0000 UTC Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.077578 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.077582 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:35 crc kubenswrapper[4857]: E0222 00:07:35.077823 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.077855 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:35 crc kubenswrapper[4857]: E0222 00:07:35.077942 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:35 crc kubenswrapper[4857]: E0222 00:07:35.078023 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.114546 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.114601 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.114617 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.114639 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.114658 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:35Z","lastTransitionTime":"2026-02-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.216793 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.216837 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.216849 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.216868 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.216888 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:35Z","lastTransitionTime":"2026-02-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.318704 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.318760 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.318777 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.318923 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.318947 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:35Z","lastTransitionTime":"2026-02-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.422757 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.422807 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.422816 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.422832 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.422841 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:35Z","lastTransitionTime":"2026-02-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.524456 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.524491 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.524500 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.524516 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.524527 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:35Z","lastTransitionTime":"2026-02-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.627678 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.627723 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.627733 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.627747 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.627755 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:35Z","lastTransitionTime":"2026-02-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.731201 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.731281 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.731300 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.731325 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.731343 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:35Z","lastTransitionTime":"2026-02-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.834617 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.834665 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.834674 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.834689 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.834699 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:35Z","lastTransitionTime":"2026-02-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.938255 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.938326 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.938360 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.938390 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:35 crc kubenswrapper[4857]: I0222 00:07:35.938408 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:35Z","lastTransitionTime":"2026-02-22T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.027092 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 21:12:03.443844206 +0000 UTC Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.042068 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.042151 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.042163 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.042179 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.042191 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:36Z","lastTransitionTime":"2026-02-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.076571 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:36 crc kubenswrapper[4857]: E0222 00:07:36.076800 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.146150 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.146228 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.146255 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.146285 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.146309 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:36Z","lastTransitionTime":"2026-02-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.249803 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.249879 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.249902 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.249932 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.249958 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:36Z","lastTransitionTime":"2026-02-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.353729 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.353773 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.353790 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.353807 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.353818 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:36Z","lastTransitionTime":"2026-02-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.457167 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.457205 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.457221 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.457243 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.457256 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:36Z","lastTransitionTime":"2026-02-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.559648 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.559705 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.559717 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.559737 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.559749 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:36Z","lastTransitionTime":"2026-02-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.662446 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.662485 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.662497 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.662512 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.662521 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:36Z","lastTransitionTime":"2026-02-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.764886 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.764935 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.764946 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.764962 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.764972 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:36Z","lastTransitionTime":"2026-02-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.866878 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.866917 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.866928 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.866942 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.866950 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:36Z","lastTransitionTime":"2026-02-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.970213 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.970315 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.970326 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.970345 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:36 crc kubenswrapper[4857]: I0222 00:07:36.970355 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:36Z","lastTransitionTime":"2026-02-22T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.028124 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 04:41:50.012140257 +0000 UTC Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.074522 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.074560 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.074568 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.074584 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.074597 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:37Z","lastTransitionTime":"2026-02-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.076828 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.076888 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:37 crc kubenswrapper[4857]: E0222 00:07:37.076957 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.077001 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:37 crc kubenswrapper[4857]: E0222 00:07:37.077156 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:37 crc kubenswrapper[4857]: E0222 00:07:37.077927 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.176913 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.176969 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.176980 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.177002 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.177015 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:37Z","lastTransitionTime":"2026-02-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.279327 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.279406 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.279423 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.279454 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.279475 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:37Z","lastTransitionTime":"2026-02-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.381740 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.381810 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.381833 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.381867 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.381889 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:37Z","lastTransitionTime":"2026-02-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.483459 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.483500 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.483511 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.483525 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.483536 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:37Z","lastTransitionTime":"2026-02-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.586271 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.586342 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.586362 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.586385 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.586404 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:37Z","lastTransitionTime":"2026-02-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.689178 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.689218 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.689231 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.689249 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.689261 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:37Z","lastTransitionTime":"2026-02-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.794475 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.794521 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.794543 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.794562 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.794574 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:37Z","lastTransitionTime":"2026-02-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.897050 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.897098 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.897111 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.897128 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.897139 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:37Z","lastTransitionTime":"2026-02-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.998956 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.998996 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.999005 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.999020 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:37 crc kubenswrapper[4857]: I0222 00:07:37.999031 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:37Z","lastTransitionTime":"2026-02-22T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.028357 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 07:46:59.232863212 +0000 UTC Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.076746 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:38 crc kubenswrapper[4857]: E0222 00:07:38.076967 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.101474 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.101507 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.101517 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.101532 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.101543 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:38Z","lastTransitionTime":"2026-02-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.203633 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.203674 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.203686 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.203699 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.203710 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:38Z","lastTransitionTime":"2026-02-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.306418 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.306485 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.306498 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.306516 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.306528 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:38Z","lastTransitionTime":"2026-02-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.409023 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.409077 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.409089 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.409106 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.409115 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:38Z","lastTransitionTime":"2026-02-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.511598 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.511635 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.511645 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.511660 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.511670 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:38Z","lastTransitionTime":"2026-02-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.614003 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.614055 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.614065 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.614079 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.614088 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:38Z","lastTransitionTime":"2026-02-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.716541 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.716578 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.716587 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.716600 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.716610 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:38Z","lastTransitionTime":"2026-02-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.819141 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.819174 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.819181 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.819195 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.819204 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:38Z","lastTransitionTime":"2026-02-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.922231 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.922268 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.922276 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.922290 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:38 crc kubenswrapper[4857]: I0222 00:07:38.922299 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:38Z","lastTransitionTime":"2026-02-22T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.024467 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.024493 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.024501 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.024514 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.024521 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:39Z","lastTransitionTime":"2026-02-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.029079 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 23:30:58.161376426 +0000 UTC Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.077586 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.077619 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:39 crc kubenswrapper[4857]: E0222 00:07:39.077758 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.077797 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:39 crc kubenswrapper[4857]: E0222 00:07:39.077893 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:39 crc kubenswrapper[4857]: E0222 00:07:39.077975 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.127243 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.127504 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.127594 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.127676 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.127746 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:39Z","lastTransitionTime":"2026-02-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.230097 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.230353 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.230422 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.230503 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.230609 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:39Z","lastTransitionTime":"2026-02-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.333925 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.333965 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.333976 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.333990 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.333999 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:39Z","lastTransitionTime":"2026-02-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.435955 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.435996 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.436005 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.436018 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.436027 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:39Z","lastTransitionTime":"2026-02-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.538396 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.538443 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.538457 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.538473 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.538482 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:39Z","lastTransitionTime":"2026-02-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.641253 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.641304 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.641316 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.641332 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.641341 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:39Z","lastTransitionTime":"2026-02-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.743657 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.743912 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.743985 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.744084 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.744172 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:39Z","lastTransitionTime":"2026-02-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.777220 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.777269 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.777280 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.777305 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.777319 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:39Z","lastTransitionTime":"2026-02-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:39 crc kubenswrapper[4857]: E0222 00:07:39.797538 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:39Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.801282 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.801317 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.801327 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.801340 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.801349 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:39Z","lastTransitionTime":"2026-02-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:39 crc kubenswrapper[4857]: E0222 00:07:39.815579 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:39Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.822425 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.822501 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.822522 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.822553 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.822575 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:39Z","lastTransitionTime":"2026-02-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:39 crc kubenswrapper[4857]: E0222 00:07:39.838370 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:39Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.843862 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.843917 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.843936 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.843962 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.843983 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:39Z","lastTransitionTime":"2026-02-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:39 crc kubenswrapper[4857]: E0222 00:07:39.864118 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:39Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.868658 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.868701 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.868710 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.868728 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.868741 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:39Z","lastTransitionTime":"2026-02-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:39 crc kubenswrapper[4857]: E0222 00:07:39.883680 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:39Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:39 crc kubenswrapper[4857]: E0222 00:07:39.883932 4857 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.885975 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.886028 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.886054 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.886069 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.886080 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:39Z","lastTransitionTime":"2026-02-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.989607 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.989675 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.989698 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.989727 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:39 crc kubenswrapper[4857]: I0222 00:07:39.989755 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:39Z","lastTransitionTime":"2026-02-22T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.029969 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 15:11:35.453692714 +0000 UTC Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.076674 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:40 crc kubenswrapper[4857]: E0222 00:07:40.077263 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.092239 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.092283 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.092293 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.092310 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.092322 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:40Z","lastTransitionTime":"2026-02-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.195473 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.195519 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.195531 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.195548 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.195557 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:40Z","lastTransitionTime":"2026-02-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.299948 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.300011 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.300028 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.300091 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.300109 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:40Z","lastTransitionTime":"2026-02-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.402505 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.402566 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.402577 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.402600 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.402619 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:40Z","lastTransitionTime":"2026-02-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.504663 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.504717 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.504728 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.504749 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.504762 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:40Z","lastTransitionTime":"2026-02-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.608843 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.608883 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.608898 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.608912 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.608924 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:40Z","lastTransitionTime":"2026-02-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.711750 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.711785 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.711795 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.711807 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.711816 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:40Z","lastTransitionTime":"2026-02-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.814230 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.814267 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.814278 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.814292 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.814301 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:40Z","lastTransitionTime":"2026-02-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.916704 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.916753 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.916765 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.916781 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:40 crc kubenswrapper[4857]: I0222 00:07:40.916794 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:40Z","lastTransitionTime":"2026-02-22T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.019780 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.019828 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.019840 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.019855 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.019869 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:41Z","lastTransitionTime":"2026-02-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.030304 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 01:49:21.123419979 +0000 UTC Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.076668 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.076698 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:41 crc kubenswrapper[4857]: E0222 00:07:41.076791 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.076809 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:41 crc kubenswrapper[4857]: E0222 00:07:41.077159 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:41 crc kubenswrapper[4857]: E0222 00:07:41.077241 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.077333 4857 scope.go:117] "RemoveContainer" containerID="13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2" Feb 22 00:07:41 crc kubenswrapper[4857]: E0222 00:07:41.077573 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.121983 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.122067 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.122085 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.122105 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.122152 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:41Z","lastTransitionTime":"2026-02-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.224242 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.224278 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.224287 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.224301 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.224311 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:41Z","lastTransitionTime":"2026-02-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.327138 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.327174 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.327186 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.327202 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.327226 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:41Z","lastTransitionTime":"2026-02-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.429586 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.429627 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.429638 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.429650 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.429661 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:41Z","lastTransitionTime":"2026-02-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.532528 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.532584 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.532601 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.532629 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.532643 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:41Z","lastTransitionTime":"2026-02-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.635114 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.635200 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.635226 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.635260 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.635286 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:41Z","lastTransitionTime":"2026-02-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.738911 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.738974 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.738984 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.739003 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.739016 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:41Z","lastTransitionTime":"2026-02-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.841794 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.841832 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.841845 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.841860 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.841869 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:41Z","lastTransitionTime":"2026-02-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.944857 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.944912 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.944925 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.944954 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:41 crc kubenswrapper[4857]: I0222 00:07:41.944969 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:41Z","lastTransitionTime":"2026-02-22T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.030927 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 12:14:29.158560141 +0000 UTC Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.047775 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.047813 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.047825 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.047842 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.047854 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:42Z","lastTransitionTime":"2026-02-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.077467 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:42 crc kubenswrapper[4857]: E0222 00:07:42.077658 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.150717 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.150788 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.150806 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.150830 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.150847 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:42Z","lastTransitionTime":"2026-02-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.255025 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.255122 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.255146 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.255176 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.255202 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:42Z","lastTransitionTime":"2026-02-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.357778 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.357820 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.357830 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.357847 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.357858 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:42Z","lastTransitionTime":"2026-02-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.460676 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.460715 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.460726 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.460743 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.460754 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:42Z","lastTransitionTime":"2026-02-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.562673 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.562703 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.562712 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.562725 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.562733 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:42Z","lastTransitionTime":"2026-02-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.665423 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.665465 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.665478 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.665496 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.665508 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:42Z","lastTransitionTime":"2026-02-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.767891 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.767917 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.767930 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.767945 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.767957 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:42Z","lastTransitionTime":"2026-02-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.869806 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.869861 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.869873 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.869887 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.869899 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:42Z","lastTransitionTime":"2026-02-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.971803 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.971847 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.971859 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.971876 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:42 crc kubenswrapper[4857]: I0222 00:07:42.971889 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:42Z","lastTransitionTime":"2026-02-22T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.031187 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 17:32:31.452654948 +0000 UTC Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.075670 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.075869 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.075997 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.076158 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.076268 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:43Z","lastTransitionTime":"2026-02-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.076507 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.076520 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.076509 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:43 crc kubenswrapper[4857]: E0222 00:07:43.076686 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:43 crc kubenswrapper[4857]: E0222 00:07:43.076852 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:43 crc kubenswrapper[4857]: E0222 00:07:43.076941 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.089786 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:43Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.100696 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:43Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.114763 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:43Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.135616 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:43Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.151576 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea7b4f51-11b4-4b6b-a691-519f425b19cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a6a9ee3b161774e7dffa093ffbff0c3fb4202d2ca7a0351ae11bd2d5de064c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f62ed2ecd8930d7f3e899f7891c87b94f1e117a23588b2ce65581a3bd37f225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9920d33f76b36664452e744820c9fb71956c5773ca81b750d6202ed6cf08c4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:43Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.168634 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:43Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.178621 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.178666 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.178677 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.178695 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.178707 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:43Z","lastTransitionTime":"2026-02-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.181504 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:43Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.192676 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:43Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.206834 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:43Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.217389 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:43Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.227536 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:43Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.244780 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:28Z\\\",\\\"message\\\":\\\"ft-dns/node-resolver-2vqwv openshift-multus/multus-4lmlt openshift-multus/multus-additional-cni-plugins-zd2xg openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-image-registry/node-ca-lz8x8 openshift-kube-controller-manager/kube-controller-manager-crc]\\\\nF0222 00:07:28.806948 6573 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z]\\\\nI0222 00:07:28.806951 6573 services_controller.go:451] Built service openshift-kube-scheduler-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler-operator/metrics_TC\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:43Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.256674 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:43Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.269339 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:43Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.279140 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:43Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.280729 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.280758 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.280767 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.280782 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.280792 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:43Z","lastTransitionTime":"2026-02-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.288282 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:43Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.298234 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:43Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.306903 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:43Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.352786 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs\") pod \"network-metrics-daemon-jnm7q\" (UID: \"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\") " pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:43 crc kubenswrapper[4857]: E0222 00:07:43.352892 4857 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 22 00:07:43 crc kubenswrapper[4857]: E0222 00:07:43.352939 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs podName:a0fb0fe8-ad7a-48a5-8a76-737cd5076f53 nodeName:}" failed. No retries permitted until 2026-02-22 00:08:15.352926396 +0000 UTC m=+102.991655649 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs") pod "network-metrics-daemon-jnm7q" (UID: "a0fb0fe8-ad7a-48a5-8a76-737cd5076f53") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.383207 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.383252 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.383263 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.383279 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.383291 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:43Z","lastTransitionTime":"2026-02-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.485006 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.485066 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.485076 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.485091 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.485100 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:43Z","lastTransitionTime":"2026-02-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.587533 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.587572 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.587580 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.587594 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.587604 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:43Z","lastTransitionTime":"2026-02-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.689477 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.689523 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.689534 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.689550 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.689562 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:43Z","lastTransitionTime":"2026-02-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.791866 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.791926 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.791942 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.791967 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.791984 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:43Z","lastTransitionTime":"2026-02-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.894218 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.894250 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.894258 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.894272 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.894281 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:43Z","lastTransitionTime":"2026-02-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.995727 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.995941 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.996024 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.996143 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:43 crc kubenswrapper[4857]: I0222 00:07:43.996211 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:43Z","lastTransitionTime":"2026-02-22T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.032073 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 08:21:49.925313844 +0000 UTC Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.076878 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:44 crc kubenswrapper[4857]: E0222 00:07:44.077183 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.098318 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.098362 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.098371 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.098387 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.098396 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:44Z","lastTransitionTime":"2026-02-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.201693 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.201760 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.201779 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.201807 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.201825 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:44Z","lastTransitionTime":"2026-02-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.303735 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.304011 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.304151 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.304241 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.304330 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:44Z","lastTransitionTime":"2026-02-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.407475 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.407521 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.407530 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.407549 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.407560 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:44Z","lastTransitionTime":"2026-02-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.502746 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4lmlt_9e857d8a-7289-4352-9de9-1b5d0bd21e8f/kube-multus/0.log" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.502848 4857 generic.go:334] "Generic (PLEG): container finished" podID="9e857d8a-7289-4352-9de9-1b5d0bd21e8f" containerID="f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15" exitCode=1 Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.502943 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4lmlt" event={"ID":"9e857d8a-7289-4352-9de9-1b5d0bd21e8f","Type":"ContainerDied","Data":"f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15"} Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.503986 4857 scope.go:117] "RemoveContainer" containerID="f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.509681 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.509717 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.509726 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.509747 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.509759 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:44Z","lastTransitionTime":"2026-02-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.522810 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:44Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.543506 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:44Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.557624 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:44Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.566753 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:44Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.585984 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:44Z\\\",\\\"message\\\":\\\"2026-02-22T00:06:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_41e23931-e004-43fa-afda-72d0ecd20e21\\\\n2026-02-22T00:06:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_41e23931-e004-43fa-afda-72d0ecd20e21 to /host/opt/cni/bin/\\\\n2026-02-22T00:06:59Z [verbose] multus-daemon started\\\\n2026-02-22T00:06:59Z [verbose] Readiness Indicator file check\\\\n2026-02-22T00:07:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:44Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.603422 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:44Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.612422 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.612461 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.612475 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.612496 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.612509 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:44Z","lastTransitionTime":"2026-02-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.618763 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:44Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.632871 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:44Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.646167 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:44Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.660767 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:44Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.683450 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:44Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.695864 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea7b4f51-11b4-4b6b-a691-519f425b19cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a6a9ee3b161774e7dffa093ffbff0c3fb4202d2ca7a0351ae11bd2d5de064c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f62ed2ecd8930d7f3e899f7891c87b94f1e117a23588b2ce65581a3bd37f225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9920d33f76b36664452e744820c9fb71956c5773ca81b750d6202ed6cf08c4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:44Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.712548 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:44Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.714596 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.714644 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.714654 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.714671 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.714691 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:44Z","lastTransitionTime":"2026-02-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.728866 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:44Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.740351 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:44Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.753237 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:44Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.771115 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:28Z\\\",\\\"message\\\":\\\"ft-dns/node-resolver-2vqwv openshift-multus/multus-4lmlt openshift-multus/multus-additional-cni-plugins-zd2xg openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-image-registry/node-ca-lz8x8 openshift-kube-controller-manager/kube-controller-manager-crc]\\\\nF0222 00:07:28.806948 6573 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z]\\\\nI0222 00:07:28.806951 6573 services_controller.go:451] Built service openshift-kube-scheduler-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler-operator/metrics_TC\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:44Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.785464 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:44Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.818263 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.818303 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.818313 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.818333 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.818345 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:44Z","lastTransitionTime":"2026-02-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.920904 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.920943 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.920952 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.920967 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:44 crc kubenswrapper[4857]: I0222 00:07:44.920977 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:44Z","lastTransitionTime":"2026-02-22T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.023571 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.023631 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.023651 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.023672 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.023686 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:45Z","lastTransitionTime":"2026-02-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.032873 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 10:14:39.405736666 +0000 UTC Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.076778 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:45 crc kubenswrapper[4857]: E0222 00:07:45.076910 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.077116 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.077209 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:45 crc kubenswrapper[4857]: E0222 00:07:45.077266 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:45 crc kubenswrapper[4857]: E0222 00:07:45.077207 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.125669 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.125706 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.125715 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.125730 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.125769 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:45Z","lastTransitionTime":"2026-02-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.228947 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.229017 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.229101 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.229133 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.229155 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:45Z","lastTransitionTime":"2026-02-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.331653 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.331684 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.331691 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.331704 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.331713 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:45Z","lastTransitionTime":"2026-02-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.434434 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.434510 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.434535 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.434571 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.434592 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:45Z","lastTransitionTime":"2026-02-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.509218 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4lmlt_9e857d8a-7289-4352-9de9-1b5d0bd21e8f/kube-multus/0.log" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.509275 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4lmlt" event={"ID":"9e857d8a-7289-4352-9de9-1b5d0bd21e8f","Type":"ContainerStarted","Data":"28685fd8939b71352e0be13ea9b6f38671b481eb13c5d08c79e788e8fc1c1f1c"} Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.521391 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:45Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.531560 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:45Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.538047 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.538077 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.538088 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.538106 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.538118 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:45Z","lastTransitionTime":"2026-02-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.544200 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28685fd8939b71352e0be13ea9b6f38671b481eb13c5d08c79e788e8fc1c1f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:44Z\\\",\\\"message\\\":\\\"2026-02-22T00:06:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_41e23931-e004-43fa-afda-72d0ecd20e21\\\\n2026-02-22T00:06:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_41e23931-e004-43fa-afda-72d0ecd20e21 to /host/opt/cni/bin/\\\\n2026-02-22T00:06:59Z [verbose] multus-daemon started\\\\n2026-02-22T00:06:59Z [verbose] Readiness Indicator file check\\\\n2026-02-22T00:07:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:45Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.559280 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:45Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.596916 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:45Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.620349 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:45Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.640762 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.640807 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.640817 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.640831 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.640840 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:45Z","lastTransitionTime":"2026-02-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.650016 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea7b4f51-11b4-4b6b-a691-519f425b19cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a6a9ee3b161774e7dffa093ffbff0c3fb4202d2ca7a0351ae11bd2d5de064c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f62ed2ecd8930d7f3e899f7891c87b94f1e117a23588b2ce65581a3bd37f225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9920d33f76b36664452e744820c9fb71956c5773ca81b750d6202ed6cf08c4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:45Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.662500 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:45Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.674801 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:45Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.686012 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:45Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.695955 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:45Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.707758 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:45Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.718026 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:45Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.734666 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:45Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.742856 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.742886 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.742894 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.742908 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.742917 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:45Z","lastTransitionTime":"2026-02-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.749597 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:45Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.759571 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:45Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.772885 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:45Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.789277 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:28Z\\\",\\\"message\\\":\\\"ft-dns/node-resolver-2vqwv openshift-multus/multus-4lmlt openshift-multus/multus-additional-cni-plugins-zd2xg openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-image-registry/node-ca-lz8x8 openshift-kube-controller-manager/kube-controller-manager-crc]\\\\nF0222 00:07:28.806948 6573 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z]\\\\nI0222 00:07:28.806951 6573 services_controller.go:451] Built service openshift-kube-scheduler-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler-operator/metrics_TC\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:45Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.845548 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.845593 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.845604 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.845623 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.845635 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:45Z","lastTransitionTime":"2026-02-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.947630 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.947673 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.947682 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.947695 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:45 crc kubenswrapper[4857]: I0222 00:07:45.947705 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:45Z","lastTransitionTime":"2026-02-22T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.033859 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 22:05:27.331432059 +0000 UTC Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.049528 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.049594 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.049606 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.049622 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.049634 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:46Z","lastTransitionTime":"2026-02-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.077053 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:46 crc kubenswrapper[4857]: E0222 00:07:46.077381 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.152278 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.152318 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.152326 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.152338 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.152347 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:46Z","lastTransitionTime":"2026-02-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.255221 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.255551 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.255656 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.255752 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.255841 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:46Z","lastTransitionTime":"2026-02-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.357988 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.358260 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.358397 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.358499 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.358587 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:46Z","lastTransitionTime":"2026-02-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.460887 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.460921 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.460932 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.460949 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.460962 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:46Z","lastTransitionTime":"2026-02-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.563349 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.563406 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.563433 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.563452 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.563462 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:46Z","lastTransitionTime":"2026-02-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.666735 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.666772 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.666781 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.666799 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.666810 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:46Z","lastTransitionTime":"2026-02-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.769112 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.769322 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.769429 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.769511 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.769585 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:46Z","lastTransitionTime":"2026-02-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.871441 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.871722 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.872068 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.872160 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.872221 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:46Z","lastTransitionTime":"2026-02-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.974267 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.974313 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.974321 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.974335 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:46 crc kubenswrapper[4857]: I0222 00:07:46.974347 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:46Z","lastTransitionTime":"2026-02-22T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.034682 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 17:12:56.398936814 +0000 UTC Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.076192 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.076410 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.076470 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.076529 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.076601 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:47Z","lastTransitionTime":"2026-02-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.076615 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:47 crc kubenswrapper[4857]: E0222 00:07:47.076846 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.076641 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:47 crc kubenswrapper[4857]: E0222 00:07:47.077027 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.077594 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:47 crc kubenswrapper[4857]: E0222 00:07:47.077794 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.179327 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.179375 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.179391 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.179413 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.179430 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:47Z","lastTransitionTime":"2026-02-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.282088 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.282119 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.282129 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.282144 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.282155 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:47Z","lastTransitionTime":"2026-02-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.384247 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.384280 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.384290 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.384304 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.384314 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:47Z","lastTransitionTime":"2026-02-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.487159 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.487210 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.487226 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.487247 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.487265 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:47Z","lastTransitionTime":"2026-02-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.588875 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.588907 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.588915 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.588927 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.588936 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:47Z","lastTransitionTime":"2026-02-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.690759 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.690787 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.690795 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.690807 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.690815 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:47Z","lastTransitionTime":"2026-02-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.794443 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.794501 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.794523 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.794550 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.794570 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:47Z","lastTransitionTime":"2026-02-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.897145 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.897208 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.897224 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.897622 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.897645 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:47Z","lastTransitionTime":"2026-02-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.999455 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.999503 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.999520 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:47 crc kubenswrapper[4857]: I0222 00:07:47.999542 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:47.999560 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:47Z","lastTransitionTime":"2026-02-22T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.035107 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 00:12:41.347375256 +0000 UTC Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.077469 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:48 crc kubenswrapper[4857]: E0222 00:07:48.077613 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.101376 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.101410 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.101421 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.101436 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.101449 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:48Z","lastTransitionTime":"2026-02-22T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.203309 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.203341 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.203352 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.203367 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.203378 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:48Z","lastTransitionTime":"2026-02-22T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.306365 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.306476 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.306494 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.306516 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.306535 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:48Z","lastTransitionTime":"2026-02-22T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.408939 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.409416 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.409566 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.409705 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.409855 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:48Z","lastTransitionTime":"2026-02-22T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.514713 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.514901 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.514977 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.515103 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.515135 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:48Z","lastTransitionTime":"2026-02-22T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.618609 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.618662 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.618681 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.618707 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.618725 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:48Z","lastTransitionTime":"2026-02-22T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.723006 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.723099 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.723111 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.723126 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.723202 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:48Z","lastTransitionTime":"2026-02-22T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.826265 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.826345 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.826362 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.826381 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.826393 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:48Z","lastTransitionTime":"2026-02-22T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.928399 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.928444 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.928455 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.928472 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:48 crc kubenswrapper[4857]: I0222 00:07:48.928485 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:48Z","lastTransitionTime":"2026-02-22T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.031158 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.031190 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.031199 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.031211 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.031220 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:49Z","lastTransitionTime":"2026-02-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.035610 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 20:00:20.944945728 +0000 UTC Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.077208 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.077259 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:49 crc kubenswrapper[4857]: E0222 00:07:49.077350 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.077213 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:49 crc kubenswrapper[4857]: E0222 00:07:49.077480 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:49 crc kubenswrapper[4857]: E0222 00:07:49.077632 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.134227 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.134292 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.134316 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.134339 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.134356 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:49Z","lastTransitionTime":"2026-02-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.237438 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.237498 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.237513 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.237532 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.237546 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:49Z","lastTransitionTime":"2026-02-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.340504 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.340547 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.340565 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.340586 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.340603 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:49Z","lastTransitionTime":"2026-02-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.443860 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.443936 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.443956 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.443981 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.443999 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:49Z","lastTransitionTime":"2026-02-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.546981 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.547076 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.547095 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.547120 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.547164 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:49Z","lastTransitionTime":"2026-02-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.651271 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.651311 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.651320 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.651335 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.651345 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:49Z","lastTransitionTime":"2026-02-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.754374 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.754414 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.754425 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.754440 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.754451 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:49Z","lastTransitionTime":"2026-02-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.857600 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.857700 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.857773 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.857802 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.857875 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:49Z","lastTransitionTime":"2026-02-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.960127 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.960195 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.960215 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.960243 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:49 crc kubenswrapper[4857]: I0222 00:07:49.960263 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:49Z","lastTransitionTime":"2026-02-22T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.036765 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 05:22:58.51646682 +0000 UTC Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.063140 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.063183 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.063194 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.063210 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.063223 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:50Z","lastTransitionTime":"2026-02-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.076755 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:50 crc kubenswrapper[4857]: E0222 00:07:50.076943 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.165170 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.165211 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.165222 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.165237 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.165251 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:50Z","lastTransitionTime":"2026-02-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.174612 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.174703 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.174731 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.174762 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.174785 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:50Z","lastTransitionTime":"2026-02-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:50 crc kubenswrapper[4857]: E0222 00:07:50.189255 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:50Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.193386 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.193456 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.193473 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.193491 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.193504 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:50Z","lastTransitionTime":"2026-02-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:50 crc kubenswrapper[4857]: E0222 00:07:50.205786 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:50Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.209862 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.209910 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.209922 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.209939 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.209953 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:50Z","lastTransitionTime":"2026-02-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:50 crc kubenswrapper[4857]: E0222 00:07:50.221900 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:50Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.225305 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.225366 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.225378 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.225393 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.225406 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:50Z","lastTransitionTime":"2026-02-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:50 crc kubenswrapper[4857]: E0222 00:07:50.238076 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:50Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.240866 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.240894 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.240903 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.240931 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.240940 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:50Z","lastTransitionTime":"2026-02-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:50 crc kubenswrapper[4857]: E0222 00:07:50.259759 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:50Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:50 crc kubenswrapper[4857]: E0222 00:07:50.259897 4857 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.267840 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.267874 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.267882 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.267899 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.267908 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:50Z","lastTransitionTime":"2026-02-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.370145 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.370186 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.370202 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.370217 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.370242 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:50Z","lastTransitionTime":"2026-02-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.472893 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.472934 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.472946 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.472963 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.472981 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:50Z","lastTransitionTime":"2026-02-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.575892 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.575939 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.575947 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.575965 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.575976 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:50Z","lastTransitionTime":"2026-02-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.678543 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.678593 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.678605 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.678624 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.678636 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:50Z","lastTransitionTime":"2026-02-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.780894 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.780935 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.780949 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.780966 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.780979 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:50Z","lastTransitionTime":"2026-02-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.883803 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.883854 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.883866 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.883883 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.883895 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:50Z","lastTransitionTime":"2026-02-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.986539 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.986595 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.986609 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.986631 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:50 crc kubenswrapper[4857]: I0222 00:07:50.986647 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:50Z","lastTransitionTime":"2026-02-22T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.037867 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 17:05:10.351398149 +0000 UTC Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.077291 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.077370 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.077434 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:51 crc kubenswrapper[4857]: E0222 00:07:51.077448 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:51 crc kubenswrapper[4857]: E0222 00:07:51.077644 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:51 crc kubenswrapper[4857]: E0222 00:07:51.077727 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.088892 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.088924 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.088932 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.088944 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.088952 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:51Z","lastTransitionTime":"2026-02-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.191679 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.191714 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.191723 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.191735 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.191745 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:51Z","lastTransitionTime":"2026-02-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.293411 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.293470 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.293480 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.293493 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.293502 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:51Z","lastTransitionTime":"2026-02-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.396903 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.397019 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.397109 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.397144 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.397166 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:51Z","lastTransitionTime":"2026-02-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.499760 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.499851 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.499870 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.499896 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.499913 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:51Z","lastTransitionTime":"2026-02-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.602509 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.602547 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.602557 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.602572 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.602583 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:51Z","lastTransitionTime":"2026-02-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.705095 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.705146 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.705159 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.705177 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.705188 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:51Z","lastTransitionTime":"2026-02-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.808373 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.808418 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.808429 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.808447 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.808559 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:51Z","lastTransitionTime":"2026-02-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.910701 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.910749 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.910758 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.910772 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:51 crc kubenswrapper[4857]: I0222 00:07:51.910780 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:51Z","lastTransitionTime":"2026-02-22T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.013535 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.013569 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.013578 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.013590 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.013599 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:52Z","lastTransitionTime":"2026-02-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.038075 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 18:00:40.971564435 +0000 UTC Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.076490 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:52 crc kubenswrapper[4857]: E0222 00:07:52.076611 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.077296 4857 scope.go:117] "RemoveContainer" containerID="13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.088795 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.115876 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.115915 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.115931 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.115946 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.115956 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:52Z","lastTransitionTime":"2026-02-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.218564 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.218603 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.218615 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.218630 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.218642 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:52Z","lastTransitionTime":"2026-02-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.320437 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.320472 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.320480 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.320493 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.320502 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:52Z","lastTransitionTime":"2026-02-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.422793 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.422842 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.422858 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.422873 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.422883 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:52Z","lastTransitionTime":"2026-02-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.525953 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.525993 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.526003 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.526018 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.526027 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:52Z","lastTransitionTime":"2026-02-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.541194 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovnkube-controller/2.log" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.544579 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerStarted","Data":"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715"} Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.545666 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.558321 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.576511 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:28Z\\\",\\\"message\\\":\\\"ft-dns/node-resolver-2vqwv openshift-multus/multus-4lmlt openshift-multus/multus-additional-cni-plugins-zd2xg openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-image-registry/node-ca-lz8x8 openshift-kube-controller-manager/kube-controller-manager-crc]\\\\nF0222 00:07:28.806948 6573 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z]\\\\nI0222 00:07:28.806951 6573 services_controller.go:451] Built service openshift-kube-scheduler-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler-operator/metrics_TC\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.587330 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7664232-712b-4a4c-bc3e-122545668834\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3453545318207b5b5c34f8ce00f1d50b9d33674903fb173c2506fc5332a21eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c1063303cf81fdf58d084698736e9405e23a7fe947708b457c67ad8677fcf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c1063303cf81fdf58d084698736e9405e23a7fe947708b457c67ad8677fcf11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.597312 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.605770 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.617604 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28685fd8939b71352e0be13ea9b6f38671b481eb13c5d08c79e788e8fc1c1f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:44Z\\\",\\\"message\\\":\\\"2026-02-22T00:06:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_41e23931-e004-43fa-afda-72d0ecd20e21\\\\n2026-02-22T00:06:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_41e23931-e004-43fa-afda-72d0ecd20e21 to /host/opt/cni/bin/\\\\n2026-02-22T00:06:59Z [verbose] multus-daemon started\\\\n2026-02-22T00:06:59Z [verbose] Readiness Indicator file check\\\\n2026-02-22T00:07:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.628468 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.628515 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.628524 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.628539 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.628550 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:52Z","lastTransitionTime":"2026-02-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.629569 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.643615 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.657908 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.669439 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea7b4f51-11b4-4b6b-a691-519f425b19cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a6a9ee3b161774e7dffa093ffbff0c3fb4202d2ca7a0351ae11bd2d5de064c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f62ed2ecd8930d7f3e899f7891c87b94f1e117a23588b2ce65581a3bd37f225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9920d33f76b36664452e744820c9fb71956c5773ca81b750d6202ed6cf08c4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.684123 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.695452 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.706060 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.751973 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.752008 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.752015 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.752029 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.752060 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:52Z","lastTransitionTime":"2026-02-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.754663 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.769290 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.780058 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.798775 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.811685 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.821106 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.854028 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.854088 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.854103 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.854120 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.854131 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:52Z","lastTransitionTime":"2026-02-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.955583 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.955614 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.955624 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.955637 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:52 crc kubenswrapper[4857]: I0222 00:07:52.955646 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:52Z","lastTransitionTime":"2026-02-22T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.038508 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 20:35:36.464530054 +0000 UTC Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.057540 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.057592 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.057604 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.057622 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.057640 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:53Z","lastTransitionTime":"2026-02-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.076885 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.076930 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:53 crc kubenswrapper[4857]: E0222 00:07:53.077005 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:53 crc kubenswrapper[4857]: E0222 00:07:53.077130 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.077149 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:53 crc kubenswrapper[4857]: E0222 00:07:53.077238 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.087738 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7664232-712b-4a4c-bc3e-122545668834\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3453545318207b5b5c34f8ce00f1d50b9d33674903fb173c2506fc5332a21eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c1063303cf81fdf58d084698736e9405e23a7fe947708b457c67ad8677fcf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c1063303cf81fdf58d084698736e9405e23a7fe947708b457c67ad8677fcf11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.098622 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.110340 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.121777 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.130419 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.141347 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28685fd8939b71352e0be13ea9b6f38671b481eb13c5d08c79e788e8fc1c1f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:44Z\\\",\\\"message\\\":\\\"2026-02-22T00:06:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_41e23931-e004-43fa-afda-72d0ecd20e21\\\\n2026-02-22T00:06:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_41e23931-e004-43fa-afda-72d0ecd20e21 to /host/opt/cni/bin/\\\\n2026-02-22T00:06:59Z [verbose] multus-daemon started\\\\n2026-02-22T00:06:59Z [verbose] Readiness Indicator file check\\\\n2026-02-22T00:07:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.152343 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.161706 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.161737 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.161746 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.161759 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.161769 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:53Z","lastTransitionTime":"2026-02-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.168209 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.187177 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.198225 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea7b4f51-11b4-4b6b-a691-519f425b19cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a6a9ee3b161774e7dffa093ffbff0c3fb4202d2ca7a0351ae11bd2d5de064c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f62ed2ecd8930d7f3e899f7891c87b94f1e117a23588b2ce65581a3bd37f225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9920d33f76b36664452e744820c9fb71956c5773ca81b750d6202ed6cf08c4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.210501 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.223284 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.233531 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.244612 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.255477 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.263740 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.263776 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.263787 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.263803 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.263812 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:53Z","lastTransitionTime":"2026-02-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.270138 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.280526 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.291433 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.308930 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:28Z\\\",\\\"message\\\":\\\"ft-dns/node-resolver-2vqwv openshift-multus/multus-4lmlt openshift-multus/multus-additional-cni-plugins-zd2xg openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-image-registry/node-ca-lz8x8 openshift-kube-controller-manager/kube-controller-manager-crc]\\\\nF0222 00:07:28.806948 6573 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z]\\\\nI0222 00:07:28.806951 6573 services_controller.go:451] Built service openshift-kube-scheduler-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler-operator/metrics_TC\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.365911 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.365947 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.365957 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.365971 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.365980 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:53Z","lastTransitionTime":"2026-02-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.467772 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.468017 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.468100 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.468163 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.468238 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:53Z","lastTransitionTime":"2026-02-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.548647 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovnkube-controller/3.log" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.549210 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovnkube-controller/2.log" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.551603 4857 generic.go:334] "Generic (PLEG): container finished" podID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerID="f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715" exitCode=1 Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.551716 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerDied","Data":"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715"} Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.551823 4857 scope.go:117] "RemoveContainer" containerID="13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.552232 4857 scope.go:117] "RemoveContainer" containerID="f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715" Feb 22 00:07:53 crc kubenswrapper[4857]: E0222 00:07:53.552467 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.563635 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7664232-712b-4a4c-bc3e-122545668834\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3453545318207b5b5c34f8ce00f1d50b9d33674903fb173c2506fc5332a21eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c1063303cf81fdf58d084698736e9405e23a7fe947708b457c67ad8677fcf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c1063303cf81fdf58d084698736e9405e23a7fe947708b457c67ad8677fcf11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.571278 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.571311 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.571321 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.571338 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.571348 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:53Z","lastTransitionTime":"2026-02-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.575663 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.588548 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.599829 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.608161 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.620421 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28685fd8939b71352e0be13ea9b6f38671b481eb13c5d08c79e788e8fc1c1f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:44Z\\\",\\\"message\\\":\\\"2026-02-22T00:06:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_41e23931-e004-43fa-afda-72d0ecd20e21\\\\n2026-02-22T00:06:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_41e23931-e004-43fa-afda-72d0ecd20e21 to /host/opt/cni/bin/\\\\n2026-02-22T00:06:59Z [verbose] multus-daemon started\\\\n2026-02-22T00:06:59Z [verbose] Readiness Indicator file check\\\\n2026-02-22T00:07:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.631061 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.643721 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.664791 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.673852 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.673900 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.673915 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.673935 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.673947 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:53Z","lastTransitionTime":"2026-02-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.675922 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea7b4f51-11b4-4b6b-a691-519f425b19cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a6a9ee3b161774e7dffa093ffbff0c3fb4202d2ca7a0351ae11bd2d5de064c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f62ed2ecd8930d7f3e899f7891c87b94f1e117a23588b2ce65581a3bd37f225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9920d33f76b36664452e744820c9fb71956c5773ca81b750d6202ed6cf08c4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.687178 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.697131 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.707357 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.716425 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.725404 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.736757 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.750009 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.761096 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.776599 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.776636 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.776648 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.776663 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.776674 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:53Z","lastTransitionTime":"2026-02-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.776620 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13df16e70680ba40dcb82de0782cf92c136bd1a1a101dfc1ee442ddddb461da2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:28Z\\\",\\\"message\\\":\\\"ft-dns/node-resolver-2vqwv openshift-multus/multus-4lmlt openshift-multus/multus-additional-cni-plugins-zd2xg openshift-network-console/networking-console-plugin-85b44fc459-gdk6g openshift-image-registry/node-ca-lz8x8 openshift-kube-controller-manager/kube-controller-manager-crc]\\\\nF0222 00:07:28.806948 6573 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:28Z is after 2025-08-24T17:21:41Z]\\\\nI0222 00:07:28.806951 6573 services_controller.go:451] Built service openshift-kube-scheduler-operator/metrics cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-kube-scheduler-operator/metrics_TC\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:52Z\\\",\\\"message\\\":\\\"ontroller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z]\\\\nI0222 00:07:52.909409 6935 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:53Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.879211 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.879263 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.879272 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.879289 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.879298 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:53Z","lastTransitionTime":"2026-02-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.981855 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.982338 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.982346 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.982359 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:53 crc kubenswrapper[4857]: I0222 00:07:53.982367 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:53Z","lastTransitionTime":"2026-02-22T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.039485 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 04:15:45.467123527 +0000 UTC Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.077497 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:54 crc kubenswrapper[4857]: E0222 00:07:54.077624 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.084109 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.084142 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.084160 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.084178 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.084191 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:54Z","lastTransitionTime":"2026-02-22T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.186283 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.186336 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.186376 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.186397 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.186408 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:54Z","lastTransitionTime":"2026-02-22T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.288455 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.288502 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.288513 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.288530 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.288541 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:54Z","lastTransitionTime":"2026-02-22T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.391453 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.391586 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.391604 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.391620 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.391631 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:54Z","lastTransitionTime":"2026-02-22T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.493457 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.493493 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.493508 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.493524 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.493534 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:54Z","lastTransitionTime":"2026-02-22T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.555354 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovnkube-controller/3.log" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.558709 4857 scope.go:117] "RemoveContainer" containerID="f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715" Feb 22 00:07:54 crc kubenswrapper[4857]: E0222 00:07:54.558850 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.568595 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7664232-712b-4a4c-bc3e-122545668834\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3453545318207b5b5c34f8ce00f1d50b9d33674903fb173c2506fc5332a21eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c1063303cf81fdf58d084698736e9405e23a7fe947708b457c67ad8677fcf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c1063303cf81fdf58d084698736e9405e23a7fe947708b457c67ad8677fcf11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.579012 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.587756 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.595423 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.595460 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.595472 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.595488 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.595498 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:54Z","lastTransitionTime":"2026-02-22T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.599638 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.610447 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.618877 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.629307 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28685fd8939b71352e0be13ea9b6f38671b481eb13c5d08c79e788e8fc1c1f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:44Z\\\",\\\"message\\\":\\\"2026-02-22T00:06:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_41e23931-e004-43fa-afda-72d0ecd20e21\\\\n2026-02-22T00:06:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_41e23931-e004-43fa-afda-72d0ecd20e21 to /host/opt/cni/bin/\\\\n2026-02-22T00:06:59Z [verbose] multus-daemon started\\\\n2026-02-22T00:06:59Z [verbose] Readiness Indicator file check\\\\n2026-02-22T00:07:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.639943 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.650543 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.664767 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.674655 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.684607 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.697122 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.697158 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.697170 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.697185 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.697202 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:54Z","lastTransitionTime":"2026-02-22T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.700819 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.711277 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea7b4f51-11b4-4b6b-a691-519f425b19cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a6a9ee3b161774e7dffa093ffbff0c3fb4202d2ca7a0351ae11bd2d5de064c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f62ed2ecd8930d7f3e899f7891c87b94f1e117a23588b2ce65581a3bd37f225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9920d33f76b36664452e744820c9fb71956c5773ca81b750d6202ed6cf08c4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.723295 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.735849 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.745178 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.754732 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.770301 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:52Z\\\",\\\"message\\\":\\\"ontroller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z]\\\\nI0222 00:07:52.909409 6935 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:54Z is after 2025-08-24T17:21:41Z" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.798949 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.798970 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.798977 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.798990 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.798999 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:54Z","lastTransitionTime":"2026-02-22T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.901629 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.901688 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.901711 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.901739 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:54 crc kubenswrapper[4857]: I0222 00:07:54.901759 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:54Z","lastTransitionTime":"2026-02-22T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.003804 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.003849 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.003863 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.003879 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.003889 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:55Z","lastTransitionTime":"2026-02-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.040400 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 03:03:10.243929229 +0000 UTC Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.077291 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.077344 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.077512 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:55 crc kubenswrapper[4857]: E0222 00:07:55.077499 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:55 crc kubenswrapper[4857]: E0222 00:07:55.077623 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:55 crc kubenswrapper[4857]: E0222 00:07:55.077785 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.106199 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.106245 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.106258 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.106282 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.106299 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:55Z","lastTransitionTime":"2026-02-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.209484 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.209542 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.209559 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.209586 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.209604 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:55Z","lastTransitionTime":"2026-02-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.311823 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.311921 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.311953 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.311984 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.312004 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:55Z","lastTransitionTime":"2026-02-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.414082 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.414115 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.414123 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.414137 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.414145 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:55Z","lastTransitionTime":"2026-02-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.517248 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.517327 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.517345 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.517373 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.517393 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:55Z","lastTransitionTime":"2026-02-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.620168 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.620255 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.620288 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.620320 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.620342 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:55Z","lastTransitionTime":"2026-02-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.722807 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.722847 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.722858 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.722874 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.722885 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:55Z","lastTransitionTime":"2026-02-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.825804 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.825871 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.825891 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.825917 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.825935 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:55Z","lastTransitionTime":"2026-02-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.928882 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.928944 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.928961 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.928985 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:55 crc kubenswrapper[4857]: I0222 00:07:55.929002 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:55Z","lastTransitionTime":"2026-02-22T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.032071 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.032136 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.032152 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.032176 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.032196 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:56Z","lastTransitionTime":"2026-02-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.041614 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 12:42:48.281204756 +0000 UTC Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.077122 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:56 crc kubenswrapper[4857]: E0222 00:07:56.077344 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.135114 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.135184 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.135203 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.135227 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.135243 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:56Z","lastTransitionTime":"2026-02-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.237788 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.237845 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.237869 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.237891 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.237909 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:56Z","lastTransitionTime":"2026-02-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.340895 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.340955 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.340975 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.341073 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.341099 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:56Z","lastTransitionTime":"2026-02-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.445320 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.445379 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.445391 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.445416 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.445429 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:56Z","lastTransitionTime":"2026-02-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.548611 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.548673 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.548685 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.548706 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.548719 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:56Z","lastTransitionTime":"2026-02-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.651574 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.651610 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.651618 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.651632 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.651641 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:56Z","lastTransitionTime":"2026-02-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.753951 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.753993 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.754008 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.754025 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.754051 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:56Z","lastTransitionTime":"2026-02-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.857216 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.857271 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.857281 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.857294 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.857304 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:56Z","lastTransitionTime":"2026-02-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.881664 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:07:56 crc kubenswrapper[4857]: E0222 00:07:56.881864 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.881828549 +0000 UTC m=+148.520557972 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.881837 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.881987 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:56 crc kubenswrapper[4857]: E0222 00:07:56.882118 4857 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 22 00:07:56 crc kubenswrapper[4857]: E0222 00:07:56.882186 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.882170589 +0000 UTC m=+148.520899842 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 22 00:07:56 crc kubenswrapper[4857]: E0222 00:07:56.882263 4857 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 22 00:07:56 crc kubenswrapper[4857]: E0222 00:07:56.882392 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.882367033 +0000 UTC m=+148.521096476 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.961849 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.961945 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.961965 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.961999 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.962021 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:56Z","lastTransitionTime":"2026-02-22T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.983029 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:56 crc kubenswrapper[4857]: I0222 00:07:56.983186 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:56 crc kubenswrapper[4857]: E0222 00:07:56.983450 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 22 00:07:56 crc kubenswrapper[4857]: E0222 00:07:56.983506 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 22 00:07:56 crc kubenswrapper[4857]: E0222 00:07:56.983521 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 22 00:07:56 crc kubenswrapper[4857]: E0222 00:07:56.983540 4857 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 22 00:07:56 crc kubenswrapper[4857]: E0222 00:07:56.983554 4857 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:07:56 crc kubenswrapper[4857]: E0222 00:07:56.983565 4857 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:07:56 crc kubenswrapper[4857]: E0222 00:07:56.983672 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.983637663 +0000 UTC m=+148.622367076 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:07:56 crc kubenswrapper[4857]: E0222 00:07:56.983719 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.983700515 +0000 UTC m=+148.622429798 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.042465 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 01:08:20.292698955 +0000 UTC Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.065849 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.065911 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.065930 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.065957 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.065980 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:57Z","lastTransitionTime":"2026-02-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.076925 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.077002 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:57 crc kubenswrapper[4857]: E0222 00:07:57.077073 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:57 crc kubenswrapper[4857]: E0222 00:07:57.077211 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.077304 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:57 crc kubenswrapper[4857]: E0222 00:07:57.077565 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.169216 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.169274 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.169292 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.169315 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.169334 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:57Z","lastTransitionTime":"2026-02-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.280675 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.280751 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.280772 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.280807 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.280827 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:57Z","lastTransitionTime":"2026-02-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.383723 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.383796 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.383816 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.383842 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.383863 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:57Z","lastTransitionTime":"2026-02-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.487719 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.487798 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.487822 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.487847 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.487869 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:57Z","lastTransitionTime":"2026-02-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.589865 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.589917 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.589929 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.589947 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.589960 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:57Z","lastTransitionTime":"2026-02-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.692779 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.692817 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.692826 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.692840 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.692851 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:57Z","lastTransitionTime":"2026-02-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.795773 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.795821 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.795836 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.795857 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.795875 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:57Z","lastTransitionTime":"2026-02-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.898426 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.898474 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.898494 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.898517 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:57 crc kubenswrapper[4857]: I0222 00:07:57.898531 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:57Z","lastTransitionTime":"2026-02-22T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.001539 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.001586 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.001599 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.001614 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.001625 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:58Z","lastTransitionTime":"2026-02-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.042919 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 04:39:37.881058495 +0000 UTC Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.077237 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:07:58 crc kubenswrapper[4857]: E0222 00:07:58.077391 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.104069 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.104104 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.104116 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.104131 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.104142 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:58Z","lastTransitionTime":"2026-02-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.206692 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.206764 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.206806 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.206841 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.206868 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:58Z","lastTransitionTime":"2026-02-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.310227 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.310303 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.310328 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.310358 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.310383 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:58Z","lastTransitionTime":"2026-02-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.413635 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.413688 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.413699 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.413717 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.413729 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:58Z","lastTransitionTime":"2026-02-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.518192 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.518234 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.518245 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.518264 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.518275 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:58Z","lastTransitionTime":"2026-02-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.621449 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.621522 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.621538 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.621556 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.621567 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:58Z","lastTransitionTime":"2026-02-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.724622 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.724698 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.724734 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.724768 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.724791 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:58Z","lastTransitionTime":"2026-02-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.827738 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.827816 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.827862 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.827899 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.827925 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:58Z","lastTransitionTime":"2026-02-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.930409 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.930460 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.930471 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.930495 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:58 crc kubenswrapper[4857]: I0222 00:07:58.930509 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:58Z","lastTransitionTime":"2026-02-22T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.034217 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.034307 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.034330 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.034361 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.034384 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:59Z","lastTransitionTime":"2026-02-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.043632 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 14:43:34.348374333 +0000 UTC Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.077130 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.077156 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.077200 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:07:59 crc kubenswrapper[4857]: E0222 00:07:59.077313 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:07:59 crc kubenswrapper[4857]: E0222 00:07:59.077448 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:07:59 crc kubenswrapper[4857]: E0222 00:07:59.077612 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.136947 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.136985 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.136996 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.137013 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.137025 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:59Z","lastTransitionTime":"2026-02-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.239575 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.239637 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.239657 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.239682 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.239703 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:59Z","lastTransitionTime":"2026-02-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.342521 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.342595 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.342617 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.342644 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.342664 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:59Z","lastTransitionTime":"2026-02-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.445458 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.445534 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.445559 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.445592 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.445618 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:59Z","lastTransitionTime":"2026-02-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.549214 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.549283 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.549300 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.549330 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.549350 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:59Z","lastTransitionTime":"2026-02-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.652376 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.652449 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.652471 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.652498 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.652515 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:59Z","lastTransitionTime":"2026-02-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.755468 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.755542 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.755565 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.755593 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.755611 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:59Z","lastTransitionTime":"2026-02-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.858568 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.858602 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.858613 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.858628 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.858640 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:59Z","lastTransitionTime":"2026-02-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.961251 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.961289 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.961300 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.961316 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:07:59 crc kubenswrapper[4857]: I0222 00:07:59.961327 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:07:59Z","lastTransitionTime":"2026-02-22T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.044302 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 14:58:46.494596674 +0000 UTC Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.064309 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.064388 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.064411 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.064440 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.064463 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:00Z","lastTransitionTime":"2026-02-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.076884 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:00 crc kubenswrapper[4857]: E0222 00:08:00.077114 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.166840 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.166902 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.166919 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.166939 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.166956 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:00Z","lastTransitionTime":"2026-02-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.269665 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.269697 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.269709 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.269726 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.269737 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:00Z","lastTransitionTime":"2026-02-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.373481 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.373541 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.373559 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.373582 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.373598 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:00Z","lastTransitionTime":"2026-02-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.462708 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.462776 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.462800 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.462831 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.462854 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:00Z","lastTransitionTime":"2026-02-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:00 crc kubenswrapper[4857]: E0222 00:08:00.486874 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.492692 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.492755 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.492774 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.492800 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.492819 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:00Z","lastTransitionTime":"2026-02-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:00 crc kubenswrapper[4857]: E0222 00:08:00.514515 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.518687 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.518743 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.518763 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.518787 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.518804 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:00Z","lastTransitionTime":"2026-02-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:00 crc kubenswrapper[4857]: E0222 00:08:00.539451 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.543376 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.543422 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.543434 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.543454 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.543468 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:00Z","lastTransitionTime":"2026-02-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:00 crc kubenswrapper[4857]: E0222 00:08:00.557353 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.561716 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.561773 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.561820 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.561840 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.561850 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:00Z","lastTransitionTime":"2026-02-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:00 crc kubenswrapper[4857]: E0222 00:08:00.578363 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:00Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:00 crc kubenswrapper[4857]: E0222 00:08:00.578563 4857 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.580343 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.580391 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.580407 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.580429 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.580445 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:00Z","lastTransitionTime":"2026-02-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.682728 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.682755 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.682763 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.682775 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.682784 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:00Z","lastTransitionTime":"2026-02-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.785422 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.785778 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.785793 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.785809 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.785821 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:00Z","lastTransitionTime":"2026-02-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.888934 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.888966 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.888975 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.888994 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.889004 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:00Z","lastTransitionTime":"2026-02-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.991386 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.991423 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.991434 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.991449 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:00 crc kubenswrapper[4857]: I0222 00:08:00.991460 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:00Z","lastTransitionTime":"2026-02-22T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.044653 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 23:31:15.352218702 +0000 UTC Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.077125 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.077196 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:01 crc kubenswrapper[4857]: E0222 00:08:01.077323 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.077342 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:01 crc kubenswrapper[4857]: E0222 00:08:01.077451 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:01 crc kubenswrapper[4857]: E0222 00:08:01.077584 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.095352 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.095406 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.095423 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.095462 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.095485 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:01Z","lastTransitionTime":"2026-02-22T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.198312 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.198361 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.198372 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.198390 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.198404 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:01Z","lastTransitionTime":"2026-02-22T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.301554 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.301592 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.301607 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.301623 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.301633 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:01Z","lastTransitionTime":"2026-02-22T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.404376 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.404426 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.404443 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.404469 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.404486 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:01Z","lastTransitionTime":"2026-02-22T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.511993 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.512080 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.512122 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.512158 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.512183 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:01Z","lastTransitionTime":"2026-02-22T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.615067 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.615128 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.615145 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.615171 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.615193 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:01Z","lastTransitionTime":"2026-02-22T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.718236 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.718295 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.718313 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.718342 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.718466 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:01Z","lastTransitionTime":"2026-02-22T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.821915 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.821995 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.822018 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.822095 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.822129 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:01Z","lastTransitionTime":"2026-02-22T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.925221 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.925307 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.925324 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.925344 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:01 crc kubenswrapper[4857]: I0222 00:08:01.925356 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:01Z","lastTransitionTime":"2026-02-22T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.029392 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.029482 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.029509 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.029610 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.029639 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:02Z","lastTransitionTime":"2026-02-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.045266 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 08:17:11.915895736 +0000 UTC Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.076571 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:02 crc kubenswrapper[4857]: E0222 00:08:02.076774 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.132790 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.132827 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.132836 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.132849 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.132859 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:02Z","lastTransitionTime":"2026-02-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.235019 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.235132 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.235153 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.235177 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.235193 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:02Z","lastTransitionTime":"2026-02-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.338169 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.338258 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.338275 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.338300 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.338316 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:02Z","lastTransitionTime":"2026-02-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.441006 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.441101 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.441124 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.441152 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.441173 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:02Z","lastTransitionTime":"2026-02-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.544377 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.544435 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.544448 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.544467 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.544479 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:02Z","lastTransitionTime":"2026-02-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.646889 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.646932 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.646941 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.646956 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.646965 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:02Z","lastTransitionTime":"2026-02-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.749827 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.749893 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.749911 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.749934 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.749953 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:02Z","lastTransitionTime":"2026-02-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.852381 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.852448 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.852464 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.852481 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.852492 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:02Z","lastTransitionTime":"2026-02-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.955788 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.955839 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.955853 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.955878 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:02 crc kubenswrapper[4857]: I0222 00:08:02.955894 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:02Z","lastTransitionTime":"2026-02-22T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.045830 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 14:42:15.99634376 +0000 UTC Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.058624 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.058679 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.058702 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.058734 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.058755 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:03Z","lastTransitionTime":"2026-02-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.077263 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.077393 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.077329 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:03 crc kubenswrapper[4857]: E0222 00:08:03.077508 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:03 crc kubenswrapper[4857]: E0222 00:08:03.077628 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:03 crc kubenswrapper[4857]: E0222 00:08:03.077771 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.092536 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.116507 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:52Z\\\",\\\"message\\\":\\\"ontroller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z]\\\\nI0222 00:07:52.909409 6935 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.129942 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7664232-712b-4a4c-bc3e-122545668834\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3453545318207b5b5c34f8ce00f1d50b9d33674903fb173c2506fc5332a21eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c1063303cf81fdf58d084698736e9405e23a7fe947708b457c67ad8677fcf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c1063303cf81fdf58d084698736e9405e23a7fe947708b457c67ad8677fcf11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.140548 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.154353 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.166430 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.166452 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.166461 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.166473 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.166482 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:03Z","lastTransitionTime":"2026-02-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.175258 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.188403 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.210891 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28685fd8939b71352e0be13ea9b6f38671b481eb13c5d08c79e788e8fc1c1f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:44Z\\\",\\\"message\\\":\\\"2026-02-22T00:06:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_41e23931-e004-43fa-afda-72d0ecd20e21\\\\n2026-02-22T00:06:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_41e23931-e004-43fa-afda-72d0ecd20e21 to /host/opt/cni/bin/\\\\n2026-02-22T00:06:59Z [verbose] multus-daemon started\\\\n2026-02-22T00:06:59Z [verbose] Readiness Indicator file check\\\\n2026-02-22T00:07:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.227816 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.243543 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.260248 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.270910 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.270991 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.271079 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.271114 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.271135 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:03Z","lastTransitionTime":"2026-02-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.277452 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.311486 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.326227 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea7b4f51-11b4-4b6b-a691-519f425b19cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a6a9ee3b161774e7dffa093ffbff0c3fb4202d2ca7a0351ae11bd2d5de064c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f62ed2ecd8930d7f3e899f7891c87b94f1e117a23588b2ce65581a3bd37f225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9920d33f76b36664452e744820c9fb71956c5773ca81b750d6202ed6cf08c4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.347947 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.366855 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.374111 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.374153 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.374164 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.374181 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.374193 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:03Z","lastTransitionTime":"2026-02-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.382496 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.402551 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.416172 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:03Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.476667 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.476740 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.476771 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.476803 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.476827 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:03Z","lastTransitionTime":"2026-02-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.579475 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.579511 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.579522 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.579538 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.579549 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:03Z","lastTransitionTime":"2026-02-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.682080 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.682119 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.682130 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.682146 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.682166 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:03Z","lastTransitionTime":"2026-02-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.785458 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.785561 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.785589 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.785620 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.785644 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:03Z","lastTransitionTime":"2026-02-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.887787 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.887855 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.887879 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.887911 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.887934 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:03Z","lastTransitionTime":"2026-02-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.991404 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.991475 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.991495 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.991522 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:03 crc kubenswrapper[4857]: I0222 00:08:03.991543 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:03Z","lastTransitionTime":"2026-02-22T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.046142 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 16:19:36.859416304 +0000 UTC Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.077143 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:04 crc kubenswrapper[4857]: E0222 00:08:04.077856 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.094358 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.094423 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.094447 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.094478 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.094501 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:04Z","lastTransitionTime":"2026-02-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.197262 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.197310 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.197322 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.197339 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.197353 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:04Z","lastTransitionTime":"2026-02-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.300093 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.300152 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.300200 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.300223 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.300238 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:04Z","lastTransitionTime":"2026-02-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.402735 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.402809 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.402832 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.402861 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.402883 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:04Z","lastTransitionTime":"2026-02-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.505749 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.505792 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.505805 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.505823 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.505834 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:04Z","lastTransitionTime":"2026-02-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.608860 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.608908 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.608923 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.608945 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.608958 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:04Z","lastTransitionTime":"2026-02-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.711585 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.711658 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.711682 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.711718 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.711741 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:04Z","lastTransitionTime":"2026-02-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.815257 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.815316 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.815333 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.815356 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.815374 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:04Z","lastTransitionTime":"2026-02-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.918472 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.918530 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.918553 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.918601 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:04 crc kubenswrapper[4857]: I0222 00:08:04.918625 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:04Z","lastTransitionTime":"2026-02-22T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.022238 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.022293 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.022310 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.022333 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.022350 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:05Z","lastTransitionTime":"2026-02-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.046759 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 18:02:29.1925642 +0000 UTC Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.076606 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.076731 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.076621 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:05 crc kubenswrapper[4857]: E0222 00:08:05.076804 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:05 crc kubenswrapper[4857]: E0222 00:08:05.076963 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:05 crc kubenswrapper[4857]: E0222 00:08:05.077350 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.124897 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.124972 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.124994 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.125022 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.125069 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:05Z","lastTransitionTime":"2026-02-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.228566 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.228641 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.228665 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.228697 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.228722 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:05Z","lastTransitionTime":"2026-02-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.332293 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.332355 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.332372 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.332404 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.332424 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:05Z","lastTransitionTime":"2026-02-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.435905 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.435965 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.435982 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.436006 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.436025 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:05Z","lastTransitionTime":"2026-02-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.538627 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.538681 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.538696 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.538719 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.538736 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:05Z","lastTransitionTime":"2026-02-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.642589 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.642670 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.642697 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.642727 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.642750 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:05Z","lastTransitionTime":"2026-02-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.745445 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.745508 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.745530 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.745556 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.745573 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:05Z","lastTransitionTime":"2026-02-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.848084 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.848143 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.848160 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.848184 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.848202 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:05Z","lastTransitionTime":"2026-02-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.951007 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.951354 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.951387 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.951416 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:05 crc kubenswrapper[4857]: I0222 00:08:05.951442 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:05Z","lastTransitionTime":"2026-02-22T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.047615 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 00:36:40.33857651 +0000 UTC Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.053732 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.053795 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.053820 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.053851 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.053871 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:06Z","lastTransitionTime":"2026-02-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.076455 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:06 crc kubenswrapper[4857]: E0222 00:08:06.076624 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.156972 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.157032 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.157096 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.157124 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.157147 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:06Z","lastTransitionTime":"2026-02-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.260321 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.260389 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.260412 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.260441 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.260465 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:06Z","lastTransitionTime":"2026-02-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.362531 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.362571 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.362584 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.362600 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.362611 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:06Z","lastTransitionTime":"2026-02-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.464573 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.464624 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.464639 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.464658 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.464672 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:06Z","lastTransitionTime":"2026-02-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.567672 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.567707 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.567715 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.567729 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.567740 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:06Z","lastTransitionTime":"2026-02-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.670269 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.670326 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.670337 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.670359 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.670373 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:06Z","lastTransitionTime":"2026-02-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.773365 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.773401 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.773415 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.773462 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.773474 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:06Z","lastTransitionTime":"2026-02-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.876503 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.876538 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.876548 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.876564 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.876574 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:06Z","lastTransitionTime":"2026-02-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.978275 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.978336 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.978354 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.978380 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:06 crc kubenswrapper[4857]: I0222 00:08:06.978397 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:06Z","lastTransitionTime":"2026-02-22T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.048102 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 14:52:38.903953293 +0000 UTC Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.077489 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.077532 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:07 crc kubenswrapper[4857]: E0222 00:08:07.077659 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.077686 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:07 crc kubenswrapper[4857]: E0222 00:08:07.077898 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:07 crc kubenswrapper[4857]: E0222 00:08:07.078085 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.080594 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.080650 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.080673 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.080701 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.080724 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:07Z","lastTransitionTime":"2026-02-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.183545 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.183586 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.183596 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.183621 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.183637 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:07Z","lastTransitionTime":"2026-02-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.324494 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.324530 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.324539 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.324564 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.324577 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:07Z","lastTransitionTime":"2026-02-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.427479 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.427510 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.427519 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.427532 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.427542 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:07Z","lastTransitionTime":"2026-02-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.531730 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.531783 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.531795 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.531814 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.531833 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:07Z","lastTransitionTime":"2026-02-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.633820 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.633857 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.633865 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.633879 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.633889 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:07Z","lastTransitionTime":"2026-02-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.736670 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.736715 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.736725 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.736740 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.736750 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:07Z","lastTransitionTime":"2026-02-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.839377 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.839416 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.839427 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.839442 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.839452 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:07Z","lastTransitionTime":"2026-02-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.941221 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.941259 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.941268 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.941285 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:07 crc kubenswrapper[4857]: I0222 00:08:07.941295 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:07Z","lastTransitionTime":"2026-02-22T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.043868 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.043911 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.043927 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.043947 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.043963 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:08Z","lastTransitionTime":"2026-02-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.049085 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 03:43:47.659179468 +0000 UTC Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.076554 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:08 crc kubenswrapper[4857]: E0222 00:08:08.076726 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.077311 4857 scope.go:117] "RemoveContainer" containerID="f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715" Feb 22 00:08:08 crc kubenswrapper[4857]: E0222 00:08:08.077455 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.146409 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.146440 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.146448 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.146461 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.146472 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:08Z","lastTransitionTime":"2026-02-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.249716 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.249766 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.249779 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.249796 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.249810 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:08Z","lastTransitionTime":"2026-02-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.352298 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.352376 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.352411 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.352440 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.352463 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:08Z","lastTransitionTime":"2026-02-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.458832 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.458908 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.458921 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.458960 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.458976 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:08Z","lastTransitionTime":"2026-02-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.561623 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.561661 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.561673 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.561690 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.561722 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:08Z","lastTransitionTime":"2026-02-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.664136 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.664184 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.664221 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.664238 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.664250 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:08Z","lastTransitionTime":"2026-02-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.767538 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.767592 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.767604 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.767621 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.767633 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:08Z","lastTransitionTime":"2026-02-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.870587 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.870654 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.870673 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.870698 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.870715 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:08Z","lastTransitionTime":"2026-02-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.974433 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.974476 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.974485 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.974500 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:08 crc kubenswrapper[4857]: I0222 00:08:08.974510 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:08Z","lastTransitionTime":"2026-02-22T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.049742 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 01:37:38.110294873 +0000 UTC Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.076691 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.076788 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.076712 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.076890 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.076900 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.076916 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.076925 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:09Z","lastTransitionTime":"2026-02-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:09 crc kubenswrapper[4857]: E0222 00:08:09.076952 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.077106 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:09 crc kubenswrapper[4857]: E0222 00:08:09.077548 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:09 crc kubenswrapper[4857]: E0222 00:08:09.077673 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.180588 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.180668 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.180687 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.180714 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.180733 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:09Z","lastTransitionTime":"2026-02-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.283313 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.283362 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.283379 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.283404 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.283422 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:09Z","lastTransitionTime":"2026-02-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.386901 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.386959 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.386978 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.387005 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.387025 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:09Z","lastTransitionTime":"2026-02-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.489768 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.489833 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.489848 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.489871 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.489887 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:09Z","lastTransitionTime":"2026-02-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.592498 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.592545 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.592565 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.592584 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.592597 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:09Z","lastTransitionTime":"2026-02-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.695082 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.695134 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.695145 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.695163 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.695174 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:09Z","lastTransitionTime":"2026-02-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.798154 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.798213 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.798230 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.798255 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.798275 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:09Z","lastTransitionTime":"2026-02-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.900789 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.900881 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.900915 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.900956 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:09 crc kubenswrapper[4857]: I0222 00:08:09.900994 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:09Z","lastTransitionTime":"2026-02-22T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.004501 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.004566 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.004584 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.004609 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.004628 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:10Z","lastTransitionTime":"2026-02-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.050334 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 17:30:41.598820861 +0000 UTC Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.076967 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:10 crc kubenswrapper[4857]: E0222 00:08:10.077187 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.108011 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.108115 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.108138 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.108167 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.108190 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:10Z","lastTransitionTime":"2026-02-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.210923 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.210983 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.211007 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.211080 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.211109 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:10Z","lastTransitionTime":"2026-02-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.314818 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.314872 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.314889 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.314913 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.314930 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:10Z","lastTransitionTime":"2026-02-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.418318 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.418414 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.418477 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.418503 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.418522 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:10Z","lastTransitionTime":"2026-02-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.522393 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.522466 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.522488 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.522515 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.522535 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:10Z","lastTransitionTime":"2026-02-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.625245 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.625317 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.625360 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.625397 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.625424 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:10Z","lastTransitionTime":"2026-02-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.729494 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.729582 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.729601 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.729626 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.729645 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:10Z","lastTransitionTime":"2026-02-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.832224 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.832335 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.832363 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.832393 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.832413 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:10Z","lastTransitionTime":"2026-02-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.913649 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.913714 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.913730 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.913755 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.913773 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:10Z","lastTransitionTime":"2026-02-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:10 crc kubenswrapper[4857]: E0222 00:08:10.933491 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.937608 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.937664 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.937684 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.937707 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.937723 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:10Z","lastTransitionTime":"2026-02-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:10 crc kubenswrapper[4857]: E0222 00:08:10.955589 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.960055 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.960086 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.960097 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.960113 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.960124 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:10Z","lastTransitionTime":"2026-02-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:10 crc kubenswrapper[4857]: E0222 00:08:10.980291 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.985009 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.985061 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.985073 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.985090 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:10 crc kubenswrapper[4857]: I0222 00:08:10.985101 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:10Z","lastTransitionTime":"2026-02-22T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:10 crc kubenswrapper[4857]: E0222 00:08:10.996977 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:10Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.002353 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.002403 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.002420 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.002447 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.002471 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:11Z","lastTransitionTime":"2026-02-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:11 crc kubenswrapper[4857]: E0222 00:08:11.021663 4857 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-22T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-22T00:08:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2a183360-3a9c-475d-a3f6-4b5064ea42f5\\\",\\\"systemUUID\\\":\\\"f5971b29-a603-4824-855e-72c4a0612518\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:11Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:11 crc kubenswrapper[4857]: E0222 00:08:11.021920 4857 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.024097 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.024158 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.024178 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.024204 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.024223 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:11Z","lastTransitionTime":"2026-02-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.050522 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 06:22:00.038680553 +0000 UTC Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.077188 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.077305 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:11 crc kubenswrapper[4857]: E0222 00:08:11.077357 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.077386 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:11 crc kubenswrapper[4857]: E0222 00:08:11.077540 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:11 crc kubenswrapper[4857]: E0222 00:08:11.077857 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.126578 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.126634 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.126651 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.126679 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.126697 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:11Z","lastTransitionTime":"2026-02-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.229926 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.229988 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.230005 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.230030 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.230072 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:11Z","lastTransitionTime":"2026-02-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.333824 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.335533 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.335543 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.335556 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.335566 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:11Z","lastTransitionTime":"2026-02-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.438244 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.438270 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.438278 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.438291 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.438300 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:11Z","lastTransitionTime":"2026-02-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.540931 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.540985 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.541004 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.541070 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.541111 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:11Z","lastTransitionTime":"2026-02-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.647694 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.647728 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.647736 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.647747 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.647756 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:11Z","lastTransitionTime":"2026-02-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.751219 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.751281 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.751305 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.751336 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.751357 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:11Z","lastTransitionTime":"2026-02-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.853796 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.853852 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.853869 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.853891 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.853907 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:11Z","lastTransitionTime":"2026-02-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.956682 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.956773 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.956797 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.956829 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:11 crc kubenswrapper[4857]: I0222 00:08:11.956848 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:11Z","lastTransitionTime":"2026-02-22T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.050871 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 23:32:26.783097908 +0000 UTC Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.060284 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.060345 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.060403 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.060432 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.060454 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:12Z","lastTransitionTime":"2026-02-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.076790 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:12 crc kubenswrapper[4857]: E0222 00:08:12.077094 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.163741 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.163790 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.163799 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.163814 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.163824 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:12Z","lastTransitionTime":"2026-02-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.267472 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.267533 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.267554 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.267582 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.267606 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:12Z","lastTransitionTime":"2026-02-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.370093 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.370176 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.370194 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.370292 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.370323 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:12Z","lastTransitionTime":"2026-02-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.472870 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.472947 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.472973 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.473006 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.473032 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:12Z","lastTransitionTime":"2026-02-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.576225 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.576262 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.576274 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.576290 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.576302 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:12Z","lastTransitionTime":"2026-02-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.679179 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.679229 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.679247 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.679272 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.679291 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:12Z","lastTransitionTime":"2026-02-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.781795 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.781855 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.781872 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.781896 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.781913 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:12Z","lastTransitionTime":"2026-02-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.885183 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.885248 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.885266 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.885292 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.885311 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:12Z","lastTransitionTime":"2026-02-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.987560 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.987629 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.987647 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.987674 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:12 crc kubenswrapper[4857]: I0222 00:08:12.987694 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:12Z","lastTransitionTime":"2026-02-22T00:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.051996 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 18:09:17.787391112 +0000 UTC Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.077549 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.077619 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:13 crc kubenswrapper[4857]: E0222 00:08:13.077718 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:13 crc kubenswrapper[4857]: E0222 00:08:13.078094 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.078178 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:13 crc kubenswrapper[4857]: E0222 00:08:13.078323 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.089829 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.089878 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.089891 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.089908 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.089922 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:13Z","lastTransitionTime":"2026-02-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.098831 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38fa10f6115a9c0cc8477195958116ab4d822975d3385d73db1c84fc62450bb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.120581 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.140167 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.157992 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-2vqwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35f87b2b-119f-455d-9a79-39dd15ac2559\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://955f14a8c5eeb477c5532a55a964758a5ab40e529b0edbbd27df2dd85309e8bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bbmn6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-2vqwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.181200 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2ed2fde8-1449-491b-ae21-2fe11ab008ae\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba8b80776a3eb950b097f39792db309d75a6a3d6a5f20efc0e7892dcf3eb2bb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m2kwn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gw6k5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.192256 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.192335 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.192359 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.192390 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.192413 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:13Z","lastTransitionTime":"2026-02-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.201265 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a5663e2-8fcd-489e-a697-97945381b881\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2639d6aee6fac7f8725687cb3bbcab6ee076f8b0e55865dde7b269b38c08162\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5044a03cb5de644af63f81b0d0056ca1277e99509f5be2ff84a6ffd9732b12e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://da6b58998d5473fc69dc77c36114e50194f280c4105173e54e852f703d8157de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.232511 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6fa6c5-4288-49de-b6bf-cc4f166c27ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3aafb4a24de2404f5b7a4b7c6b27899ff6c8d4465dd3f08e6a34cb992cba1457\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1060f1369409cd57f1da6832f4ed1a9fa32895aa1a5538309b57852090eeb3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ea86ddb49a1f2bf53d3d1da1f597449e97b6af4b220d9be1026713a5800f7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1896b6a2a654d34c7fad91b26d30d6746f767153d4e0c43fb15f6f3c973ad77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://deb8310399ddeba36f5fd4a2d53576b3494c4bb30662edd7a54c0132744d5a16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef58af9b7cc21446739f7ce22f165c01521ff3105461c4f36e132b19d4eb8024\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://135987523a7ce4fb3e65a19845828c4282acc6f3b4be3c071a8cfdca749c8211\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0554da9f5dd6337501e6f63d016cb62629efa8f12f88ddab15f088cfebe3a4a7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.248612 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea7b4f51-11b4-4b6b-a691-519f425b19cb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21a6a9ee3b161774e7dffa093ffbff0c3fb4202d2ca7a0351ae11bd2d5de064c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f62ed2ecd8930d7f3e899f7891c87b94f1e117a23588b2ce65581a3bd37f225\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9920d33f76b36664452e744820c9fb71956c5773ca81b750d6202ed6cf08c4d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03c09e694f3a080caeb29aacdeb285ad218858fad21ffb5b9d986aadcc56e711\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.269836 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fec51cf7-f11e-4829-bcd2-08ac17cdce91\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b77e89b865cfea0ee121d26e1c6d2b04e38bfbc52955cdcd40d711449f1c58ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42ebf3aafae375716c009e0b09abaa40fab0f180a43e6e23afee08d3cd01024f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c80cec892ad0a7c21fd65468a84ebc724bf943159dacd5757dda459daad80132\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2869426ea24ed3f4d0902d73cb191801e94b34712ab1473f148d514f435ad30d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22cb9d7f4ad1d156c5e8973b809203ee7d123d04bc05e607463b61c65f6f1acd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6361f263f41cb4d476f7d718b1a693b6924171b8c2c9a45514335ee2f368c2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://07c95c86992786e0fe59530404a4de0f5da21737190a75c7e9a04157812abc7e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:07:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-28jk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zd2xg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.284861 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-74d6t\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:11Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jnm7q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.295508 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.295568 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.295583 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.295614 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.295631 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:13Z","lastTransitionTime":"2026-02-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.304883 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.334785 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a0db241-7fc5-4d71-a8cc-534f4d303883\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:52Z\\\",\\\"message\\\":\\\"ontroller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:07:52Z is after 2025-08-24T17:21:41Z]\\\\nI0222 00:07:52.909409 6935 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-route-controller-manager/route-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"18746a4d-8a63-458a-b7e3-8fb89ff95fc0\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-route-controller-manager/route-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:07:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7fw59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6rc8w\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.350626 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7664232-712b-4a4c-bc3e-122545668834\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a3453545318207b5b5c34f8ce00f1d50b9d33674903fb173c2506fc5332a21eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2c1063303cf81fdf58d084698736e9405e23a7fe947708b457c67ad8677fcf11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c1063303cf81fdf58d084698736e9405e23a7fe947708b457c67ad8677fcf11\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.367417 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ac4b3894ff889a7c62644959b9dca092bd2d3f1c1bbc12b4ab5a754896934a3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.392982 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4lmlt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9e857d8a-7289-4352-9de9-1b5d0bd21e8f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://28685fd8939b71352e0be13ea9b6f38671b481eb13c5d08c79e788e8fc1c1f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-22T00:07:44Z\\\",\\\"message\\\":\\\"2026-02-22T00:06:59+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_41e23931-e004-43fa-afda-72d0ecd20e21\\\\n2026-02-22T00:06:59+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_41e23931-e004-43fa-afda-72d0ecd20e21 to /host/opt/cni/bin/\\\\n2026-02-22T00:06:59Z [verbose] multus-daemon started\\\\n2026-02-22T00:06:59Z [verbose] Readiness Indicator file check\\\\n2026-02-22T00:07:44Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kwrkb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:57Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4lmlt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.398717 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.398760 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.398771 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.398789 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.398800 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:13Z","lastTransitionTime":"2026-02-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.408893 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8763e232-ea3f-473d-92df-a087589ad717\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a983cf1f781b31fe48601c675a37f6a029fb03af05f192e6eb7470d98af2cf98\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5b0b4247dc15c349c85a71a6666c7bec0d06acf5a065a5519a26f194aa3a70d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxzd8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:07:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-cp7qh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.429520 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfed332-8a77-48dd-b3c6-7068b918dc71\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-22T00:06:53Z\\\",\\\"message\\\":\\\"le observer\\\\nW0222 00:06:52.731812 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0222 00:06:52.731933 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0222 00:06:52.732929 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3325859154/tls.crt::/tmp/serving-cert-3325859154/tls.key\\\\\\\"\\\\nI0222 00:06:53.118401 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0222 00:06:53.124579 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0222 00:06:53.124604 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0222 00:06:53.124626 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0222 00:06:53.124634 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0222 00:06:53.132085 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0222 00:06:53.132102 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0222 00:06:53.132113 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132119 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0222 00:06:53.132125 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0222 00:06:53.132130 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0222 00:06:53.132135 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0222 00:06:53.132139 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0222 00:06:53.134769 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:47Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:36Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-22T00:06:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-22T00:06:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:33Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.449264 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:54Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://519684c1b312bc141be234c1d7d8c7e4c43cd92366569228d6e7cd2bb09cabfb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3794b314e5e551eb5d8c04881bfa95eb1488e64b2bf176929fd1ad28cebd0f20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.464851 4857 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-lz8x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9642d27-573f-4a44-b37c-a636dc22439f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-22T00:06:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6869f0f72d896f6ad1fc7af26843e053f74c72b54c7587f9ada51a3dd55e79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-22T00:06:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r98zd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-22T00:06:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-lz8x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-22T00:08:13Z is after 2025-08-24T17:21:41Z" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.501675 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.501737 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.501761 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.501862 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.501920 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:13Z","lastTransitionTime":"2026-02-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.603996 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.604075 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.604094 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.604116 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.604133 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:13Z","lastTransitionTime":"2026-02-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.706818 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.706863 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.706871 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.706888 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.706898 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:13Z","lastTransitionTime":"2026-02-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.808557 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.808631 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.808659 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.808699 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.808716 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:13Z","lastTransitionTime":"2026-02-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.910696 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.910735 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.910743 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.910757 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:13 crc kubenswrapper[4857]: I0222 00:08:13.910767 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:13Z","lastTransitionTime":"2026-02-22T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.012877 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.012915 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.012927 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.012942 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.012953 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:14Z","lastTransitionTime":"2026-02-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.052878 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 06:27:28.438091142 +0000 UTC Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.077318 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:14 crc kubenswrapper[4857]: E0222 00:08:14.077542 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.116265 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.116344 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.116364 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.116391 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.116411 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:14Z","lastTransitionTime":"2026-02-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.219216 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.219278 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.219295 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.219320 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.219381 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:14Z","lastTransitionTime":"2026-02-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.321971 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.322031 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.322064 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.322084 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.322096 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:14Z","lastTransitionTime":"2026-02-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.424553 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.424585 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.424605 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.424621 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.424631 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:14Z","lastTransitionTime":"2026-02-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.526962 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.527010 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.527020 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.527056 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.527068 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:14Z","lastTransitionTime":"2026-02-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.629392 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.629426 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.629443 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.629460 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.629472 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:14Z","lastTransitionTime":"2026-02-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.731770 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.731801 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.731813 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.731828 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.731840 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:14Z","lastTransitionTime":"2026-02-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.835706 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.835776 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.835792 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.835817 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.835835 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:14Z","lastTransitionTime":"2026-02-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.938806 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.938864 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.938882 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.938906 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:14 crc kubenswrapper[4857]: I0222 00:08:14.938924 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:14Z","lastTransitionTime":"2026-02-22T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.042281 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.042351 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.042368 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.042390 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.042411 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:15Z","lastTransitionTime":"2026-02-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.053831 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 19:25:25.811582928 +0000 UTC Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.076787 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.076932 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.077268 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:15 crc kubenswrapper[4857]: E0222 00:08:15.077461 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:15 crc kubenswrapper[4857]: E0222 00:08:15.077550 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:15 crc kubenswrapper[4857]: E0222 00:08:15.077662 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.145934 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.146084 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.146116 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.146149 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.146200 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:15Z","lastTransitionTime":"2026-02-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.248766 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.248830 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.248842 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.248858 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.248869 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:15Z","lastTransitionTime":"2026-02-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.351352 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.351392 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.351405 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.351420 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.351430 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:15Z","lastTransitionTime":"2026-02-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.418370 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs\") pod \"network-metrics-daemon-jnm7q\" (UID: \"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\") " pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:15 crc kubenswrapper[4857]: E0222 00:08:15.418523 4857 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 22 00:08:15 crc kubenswrapper[4857]: E0222 00:08:15.418574 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs podName:a0fb0fe8-ad7a-48a5-8a76-737cd5076f53 nodeName:}" failed. No retries permitted until 2026-02-22 00:09:19.418560045 +0000 UTC m=+167.057289298 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs") pod "network-metrics-daemon-jnm7q" (UID: "a0fb0fe8-ad7a-48a5-8a76-737cd5076f53") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.454357 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.454385 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.454393 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.454407 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.454417 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:15Z","lastTransitionTime":"2026-02-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.557205 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.557263 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.557279 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.557300 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.557312 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:15Z","lastTransitionTime":"2026-02-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.660441 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.660499 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.660515 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.660537 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.660554 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:15Z","lastTransitionTime":"2026-02-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.764194 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.764264 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.764283 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.764310 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.764329 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:15Z","lastTransitionTime":"2026-02-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.868194 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.868256 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.868277 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.868303 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.868324 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:15Z","lastTransitionTime":"2026-02-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.971398 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.971471 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.971491 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.971523 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:15 crc kubenswrapper[4857]: I0222 00:08:15.971541 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:15Z","lastTransitionTime":"2026-02-22T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.054388 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 08:11:53.995101798 +0000 UTC Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.074627 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.074662 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.074670 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.074682 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.074690 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:16Z","lastTransitionTime":"2026-02-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.077155 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:16 crc kubenswrapper[4857]: E0222 00:08:16.077324 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.177404 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.177456 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.177470 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.177495 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.177509 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:16Z","lastTransitionTime":"2026-02-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.279829 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.279860 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.279871 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.279884 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.279893 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:16Z","lastTransitionTime":"2026-02-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.382564 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.382621 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.382638 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.382661 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.382681 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:16Z","lastTransitionTime":"2026-02-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.485520 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.485606 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.485625 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.485665 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.485683 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:16Z","lastTransitionTime":"2026-02-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.588497 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.588538 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.588551 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.588567 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.588578 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:16Z","lastTransitionTime":"2026-02-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.690799 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.690848 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.690866 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.690892 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.690913 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:16Z","lastTransitionTime":"2026-02-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.793497 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.793540 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.793550 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.793567 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.793593 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:16Z","lastTransitionTime":"2026-02-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.895968 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.896007 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.896019 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.896080 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.896094 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:16Z","lastTransitionTime":"2026-02-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.999346 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.999403 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.999418 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.999438 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:16 crc kubenswrapper[4857]: I0222 00:08:16.999452 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:16Z","lastTransitionTime":"2026-02-22T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.054631 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 22:52:09.818381803 +0000 UTC Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.077096 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.077229 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.077212 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:17 crc kubenswrapper[4857]: E0222 00:08:17.077352 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:17 crc kubenswrapper[4857]: E0222 00:08:17.077599 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:17 crc kubenswrapper[4857]: E0222 00:08:17.077769 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.102892 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.102974 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.102992 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.103022 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.103072 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:17Z","lastTransitionTime":"2026-02-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.207137 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.207225 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.207246 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.207278 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.207296 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:17Z","lastTransitionTime":"2026-02-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.311721 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.311808 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.311827 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.311857 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.311883 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:17Z","lastTransitionTime":"2026-02-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.414714 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.414800 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.414819 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.414858 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.414878 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:17Z","lastTransitionTime":"2026-02-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.517916 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.517962 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.517977 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.517996 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.518009 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:17Z","lastTransitionTime":"2026-02-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.620070 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.620112 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.620122 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.620137 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.620147 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:17Z","lastTransitionTime":"2026-02-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.723527 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.723608 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.723637 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.723683 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.723707 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:17Z","lastTransitionTime":"2026-02-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.827016 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.827140 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.827171 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.827204 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.827230 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:17Z","lastTransitionTime":"2026-02-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.930188 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.930262 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.930275 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.930295 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:17 crc kubenswrapper[4857]: I0222 00:08:17.930307 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:17Z","lastTransitionTime":"2026-02-22T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.032550 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.032591 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.032603 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.032619 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.032631 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:18Z","lastTransitionTime":"2026-02-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.055325 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 02:46:06.895622653 +0000 UTC Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.077185 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:18 crc kubenswrapper[4857]: E0222 00:08:18.077397 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.135418 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.135494 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.135514 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.135536 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.135553 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:18Z","lastTransitionTime":"2026-02-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.239351 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.239417 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.239436 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.239462 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.239489 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:18Z","lastTransitionTime":"2026-02-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.341681 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.341748 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.341765 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.341790 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.341807 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:18Z","lastTransitionTime":"2026-02-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.444887 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.444953 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.444970 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.444996 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.445013 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:18Z","lastTransitionTime":"2026-02-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.547376 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.547509 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.547545 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.547574 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.547595 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:18Z","lastTransitionTime":"2026-02-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.649966 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.650009 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.650023 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.650064 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.650078 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:18Z","lastTransitionTime":"2026-02-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.752296 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.752357 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.752378 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.752398 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.752413 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:18Z","lastTransitionTime":"2026-02-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.854860 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.854906 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.854916 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.854932 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.854944 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:18Z","lastTransitionTime":"2026-02-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.957226 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.957268 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.957279 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.957295 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:18 crc kubenswrapper[4857]: I0222 00:08:18.957307 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:18Z","lastTransitionTime":"2026-02-22T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.055704 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 14:13:00.111888688 +0000 UTC Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.059606 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.059647 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.059658 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.059675 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.059689 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:19Z","lastTransitionTime":"2026-02-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.077650 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.077696 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.077901 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:19 crc kubenswrapper[4857]: E0222 00:08:19.078003 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.078215 4857 scope.go:117] "RemoveContainer" containerID="f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715" Feb 22 00:08:19 crc kubenswrapper[4857]: E0222 00:08:19.078222 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:19 crc kubenswrapper[4857]: E0222 00:08:19.078346 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" Feb 22 00:08:19 crc kubenswrapper[4857]: E0222 00:08:19.078331 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.161347 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.161400 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.161411 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.161429 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.161441 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:19Z","lastTransitionTime":"2026-02-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.263288 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.263358 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.263369 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.263384 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.263393 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:19Z","lastTransitionTime":"2026-02-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.365930 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.365981 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.365992 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.366009 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.366018 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:19Z","lastTransitionTime":"2026-02-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.468758 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.468805 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.468815 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.468833 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.468846 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:19Z","lastTransitionTime":"2026-02-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.571982 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.572083 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.572101 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.572125 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.572142 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:19Z","lastTransitionTime":"2026-02-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.674085 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.674149 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.674162 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.674183 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.674197 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:19Z","lastTransitionTime":"2026-02-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.776524 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.776573 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.776590 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.776612 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.776623 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:19Z","lastTransitionTime":"2026-02-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.879712 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.879751 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.879762 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.879778 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.879789 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:19Z","lastTransitionTime":"2026-02-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.982726 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.982771 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.982783 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.982800 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:19 crc kubenswrapper[4857]: I0222 00:08:19.982812 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:19Z","lastTransitionTime":"2026-02-22T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.056341 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 13:07:17.769901899 +0000 UTC Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.076932 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:20 crc kubenswrapper[4857]: E0222 00:08:20.077235 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.085326 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.085366 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.085377 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.085392 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.085404 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:20Z","lastTransitionTime":"2026-02-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.186850 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.186886 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.186894 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.186907 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.186918 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:20Z","lastTransitionTime":"2026-02-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.289070 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.289120 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.289138 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.289157 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.289170 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:20Z","lastTransitionTime":"2026-02-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.391022 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.391079 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.391088 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.391103 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.391114 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:20Z","lastTransitionTime":"2026-02-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.493340 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.493395 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.493411 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.493431 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.493449 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:20Z","lastTransitionTime":"2026-02-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.595802 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.595844 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.595854 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.595871 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.595883 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:20Z","lastTransitionTime":"2026-02-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.698146 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.698190 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.698201 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.698216 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.698226 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:20Z","lastTransitionTime":"2026-02-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.800606 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.800659 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.800673 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.800690 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.800701 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:20Z","lastTransitionTime":"2026-02-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.903152 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.903196 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.903208 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.903224 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:20 crc kubenswrapper[4857]: I0222 00:08:20.903236 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:20Z","lastTransitionTime":"2026-02-22T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.005868 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.005908 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.005916 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.005928 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.005937 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:21Z","lastTransitionTime":"2026-02-22T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.056689 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 15:41:39.514638803 +0000 UTC Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.077161 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.077283 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:21 crc kubenswrapper[4857]: E0222 00:08:21.077465 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.077695 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:21 crc kubenswrapper[4857]: E0222 00:08:21.077812 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:21 crc kubenswrapper[4857]: E0222 00:08:21.078029 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.108156 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.108208 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.108226 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.108251 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.108270 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:21Z","lastTransitionTime":"2026-02-22T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.211220 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.211289 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.211307 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.211331 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.211348 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:21Z","lastTransitionTime":"2026-02-22T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.314882 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.314948 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.314964 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.314989 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.315007 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:21Z","lastTransitionTime":"2026-02-22T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.398272 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.398328 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.398344 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.398367 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.398384 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:21Z","lastTransitionTime":"2026-02-22T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.437551 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.437624 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.437645 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.437668 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.437685 4857 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-22T00:08:21Z","lastTransitionTime":"2026-02-22T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.470528 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd"] Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.471198 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.475633 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.475715 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.476484 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.477121 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.498085 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=88.4980249 podStartE2EDuration="1m28.4980249s" podCreationTimestamp="2026-02-22 00:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:21.497621699 +0000 UTC m=+109.136350992" watchObservedRunningTime="2026-02-22 00:08:21.4980249 +0000 UTC m=+109.136754183" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.531191 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-lz8x8" podStartSLOduration=85.531158722 podStartE2EDuration="1m25.531158722s" podCreationTimestamp="2026-02-22 00:06:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:21.530450952 +0000 UTC m=+109.169180255" watchObservedRunningTime="2026-02-22 00:08:21.531158722 +0000 UTC m=+109.169888015" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.568857 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-4lmlt" podStartSLOduration=85.568836162 podStartE2EDuration="1m25.568836162s" podCreationTimestamp="2026-02-22 00:06:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:21.551504995 +0000 UTC m=+109.190234258" watchObservedRunningTime="2026-02-22 00:08:21.568836162 +0000 UTC m=+109.207565425" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.569032 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-cp7qh" podStartSLOduration=84.569026058 podStartE2EDuration="1m24.569026058s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:21.568421631 +0000 UTC m=+109.207150914" watchObservedRunningTime="2026-02-22 00:08:21.569026058 +0000 UTC m=+109.207755331" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.582947 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podStartSLOduration=85.582931159 podStartE2EDuration="1m25.582931159s" podCreationTimestamp="2026-02-22 00:06:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:21.582562239 +0000 UTC m=+109.221291512" watchObservedRunningTime="2026-02-22 00:08:21.582931159 +0000 UTC m=+109.221660422" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.586238 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/361de3fd-52fe-4862-88e0-3f045fd896db-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-z2dzd\" (UID: \"361de3fd-52fe-4862-88e0-3f045fd896db\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.586311 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/361de3fd-52fe-4862-88e0-3f045fd896db-service-ca\") pod \"cluster-version-operator-5c965bbfc6-z2dzd\" (UID: \"361de3fd-52fe-4862-88e0-3f045fd896db\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.586360 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/361de3fd-52fe-4862-88e0-3f045fd896db-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-z2dzd\" (UID: \"361de3fd-52fe-4862-88e0-3f045fd896db\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.586408 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/361de3fd-52fe-4862-88e0-3f045fd896db-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-z2dzd\" (UID: \"361de3fd-52fe-4862-88e0-3f045fd896db\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.586443 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/361de3fd-52fe-4862-88e0-3f045fd896db-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-z2dzd\" (UID: \"361de3fd-52fe-4862-88e0-3f045fd896db\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.606151 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=88.606134722 podStartE2EDuration="1m28.606134722s" podCreationTimestamp="2026-02-22 00:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:21.605891205 +0000 UTC m=+109.244620478" watchObservedRunningTime="2026-02-22 00:08:21.606134722 +0000 UTC m=+109.244863985" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.637088 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=88.637068163 podStartE2EDuration="1m28.637068163s" podCreationTimestamp="2026-02-22 00:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:21.636509347 +0000 UTC m=+109.275238610" watchObservedRunningTime="2026-02-22 00:08:21.637068163 +0000 UTC m=+109.275797426" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.653460 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=59.653434493 podStartE2EDuration="59.653434493s" podCreationTimestamp="2026-02-22 00:07:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:21.652884668 +0000 UTC m=+109.291613941" watchObservedRunningTime="2026-02-22 00:08:21.653434493 +0000 UTC m=+109.292163786" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.687648 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/361de3fd-52fe-4862-88e0-3f045fd896db-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-z2dzd\" (UID: \"361de3fd-52fe-4862-88e0-3f045fd896db\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.687722 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/361de3fd-52fe-4862-88e0-3f045fd896db-service-ca\") pod \"cluster-version-operator-5c965bbfc6-z2dzd\" (UID: \"361de3fd-52fe-4862-88e0-3f045fd896db\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.687753 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/361de3fd-52fe-4862-88e0-3f045fd896db-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-z2dzd\" (UID: \"361de3fd-52fe-4862-88e0-3f045fd896db\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.687807 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/361de3fd-52fe-4862-88e0-3f045fd896db-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-z2dzd\" (UID: \"361de3fd-52fe-4862-88e0-3f045fd896db\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.687872 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/361de3fd-52fe-4862-88e0-3f045fd896db-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-z2dzd\" (UID: \"361de3fd-52fe-4862-88e0-3f045fd896db\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.687943 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/361de3fd-52fe-4862-88e0-3f045fd896db-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-z2dzd\" (UID: \"361de3fd-52fe-4862-88e0-3f045fd896db\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.688073 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/361de3fd-52fe-4862-88e0-3f045fd896db-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-z2dzd\" (UID: \"361de3fd-52fe-4862-88e0-3f045fd896db\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.689277 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/361de3fd-52fe-4862-88e0-3f045fd896db-service-ca\") pod \"cluster-version-operator-5c965bbfc6-z2dzd\" (UID: \"361de3fd-52fe-4862-88e0-3f045fd896db\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.711991 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/361de3fd-52fe-4862-88e0-3f045fd896db-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-z2dzd\" (UID: \"361de3fd-52fe-4862-88e0-3f045fd896db\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.714905 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/361de3fd-52fe-4862-88e0-3f045fd896db-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-z2dzd\" (UID: \"361de3fd-52fe-4862-88e0-3f045fd896db\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.765642 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-2vqwv" podStartSLOduration=85.76561527 podStartE2EDuration="1m25.76561527s" podCreationTimestamp="2026-02-22 00:06:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:21.761889986 +0000 UTC m=+109.400619239" watchObservedRunningTime="2026-02-22 00:08:21.76561527 +0000 UTC m=+109.404344543" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.781624 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-zd2xg" podStartSLOduration=85.78160431 podStartE2EDuration="1m25.78160431s" podCreationTimestamp="2026-02-22 00:06:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:21.781067146 +0000 UTC m=+109.419796399" watchObservedRunningTime="2026-02-22 00:08:21.78160431 +0000 UTC m=+109.420333573" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.793270 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" Feb 22 00:08:21 crc kubenswrapper[4857]: I0222 00:08:21.845176 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=29.845159479 podStartE2EDuration="29.845159479s" podCreationTimestamp="2026-02-22 00:07:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:21.845011765 +0000 UTC m=+109.483741018" watchObservedRunningTime="2026-02-22 00:08:21.845159479 +0000 UTC m=+109.483888722" Feb 22 00:08:22 crc kubenswrapper[4857]: I0222 00:08:22.057228 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 10:51:07.404946195 +0000 UTC Feb 22 00:08:22 crc kubenswrapper[4857]: I0222 00:08:22.057343 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 22 00:08:22 crc kubenswrapper[4857]: I0222 00:08:22.066499 4857 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 22 00:08:22 crc kubenswrapper[4857]: I0222 00:08:22.076848 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:22 crc kubenswrapper[4857]: E0222 00:08:22.077184 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:22 crc kubenswrapper[4857]: I0222 00:08:22.645173 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" event={"ID":"361de3fd-52fe-4862-88e0-3f045fd896db","Type":"ContainerStarted","Data":"0d390d84b5f76e4529613158201976618a39e83fa1f3cc302f72019b398a3aa2"} Feb 22 00:08:22 crc kubenswrapper[4857]: I0222 00:08:22.645244 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" event={"ID":"361de3fd-52fe-4862-88e0-3f045fd896db","Type":"ContainerStarted","Data":"f8debd9afc05df246c09941fd1ae27a5d743db576908c703a90fbc55217a78ec"} Feb 22 00:08:22 crc kubenswrapper[4857]: I0222 00:08:22.661349 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-z2dzd" podStartSLOduration=86.661311848 podStartE2EDuration="1m26.661311848s" podCreationTimestamp="2026-02-22 00:06:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:22.660761332 +0000 UTC m=+110.299490675" watchObservedRunningTime="2026-02-22 00:08:22.661311848 +0000 UTC m=+110.300041141" Feb 22 00:08:23 crc kubenswrapper[4857]: I0222 00:08:23.077174 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:23 crc kubenswrapper[4857]: I0222 00:08:23.078398 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:23 crc kubenswrapper[4857]: I0222 00:08:23.078495 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:23 crc kubenswrapper[4857]: E0222 00:08:23.078629 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:23 crc kubenswrapper[4857]: E0222 00:08:23.078750 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:23 crc kubenswrapper[4857]: E0222 00:08:23.078892 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:24 crc kubenswrapper[4857]: I0222 00:08:24.077204 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:24 crc kubenswrapper[4857]: E0222 00:08:24.077349 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:25 crc kubenswrapper[4857]: I0222 00:08:25.077160 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:25 crc kubenswrapper[4857]: E0222 00:08:25.077285 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:25 crc kubenswrapper[4857]: I0222 00:08:25.077354 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:25 crc kubenswrapper[4857]: I0222 00:08:25.077160 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:25 crc kubenswrapper[4857]: E0222 00:08:25.077500 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:25 crc kubenswrapper[4857]: E0222 00:08:25.077759 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:26 crc kubenswrapper[4857]: I0222 00:08:26.076747 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:26 crc kubenswrapper[4857]: E0222 00:08:26.076951 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:27 crc kubenswrapper[4857]: I0222 00:08:27.077162 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:27 crc kubenswrapper[4857]: I0222 00:08:27.077168 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:27 crc kubenswrapper[4857]: I0222 00:08:27.078342 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:27 crc kubenswrapper[4857]: E0222 00:08:27.078533 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:27 crc kubenswrapper[4857]: E0222 00:08:27.078639 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:27 crc kubenswrapper[4857]: E0222 00:08:27.078816 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:28 crc kubenswrapper[4857]: I0222 00:08:28.076847 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:28 crc kubenswrapper[4857]: E0222 00:08:28.076969 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:29 crc kubenswrapper[4857]: I0222 00:08:29.076621 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:29 crc kubenswrapper[4857]: E0222 00:08:29.076994 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:29 crc kubenswrapper[4857]: I0222 00:08:29.076686 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:29 crc kubenswrapper[4857]: I0222 00:08:29.076678 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:29 crc kubenswrapper[4857]: E0222 00:08:29.077121 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:29 crc kubenswrapper[4857]: E0222 00:08:29.077311 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:30 crc kubenswrapper[4857]: I0222 00:08:30.077218 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:30 crc kubenswrapper[4857]: E0222 00:08:30.077352 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:30 crc kubenswrapper[4857]: I0222 00:08:30.675930 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4lmlt_9e857d8a-7289-4352-9de9-1b5d0bd21e8f/kube-multus/1.log" Feb 22 00:08:30 crc kubenswrapper[4857]: I0222 00:08:30.676503 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4lmlt_9e857d8a-7289-4352-9de9-1b5d0bd21e8f/kube-multus/0.log" Feb 22 00:08:30 crc kubenswrapper[4857]: I0222 00:08:30.676558 4857 generic.go:334] "Generic (PLEG): container finished" podID="9e857d8a-7289-4352-9de9-1b5d0bd21e8f" containerID="28685fd8939b71352e0be13ea9b6f38671b481eb13c5d08c79e788e8fc1c1f1c" exitCode=1 Feb 22 00:08:30 crc kubenswrapper[4857]: I0222 00:08:30.676594 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4lmlt" event={"ID":"9e857d8a-7289-4352-9de9-1b5d0bd21e8f","Type":"ContainerDied","Data":"28685fd8939b71352e0be13ea9b6f38671b481eb13c5d08c79e788e8fc1c1f1c"} Feb 22 00:08:30 crc kubenswrapper[4857]: I0222 00:08:30.676658 4857 scope.go:117] "RemoveContainer" containerID="f182387772d3ad86137cb7f461c8904d3b32b93688bac6d422df08c31f1b4a15" Feb 22 00:08:30 crc kubenswrapper[4857]: I0222 00:08:30.677284 4857 scope.go:117] "RemoveContainer" containerID="28685fd8939b71352e0be13ea9b6f38671b481eb13c5d08c79e788e8fc1c1f1c" Feb 22 00:08:30 crc kubenswrapper[4857]: E0222 00:08:30.677598 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-4lmlt_openshift-multus(9e857d8a-7289-4352-9de9-1b5d0bd21e8f)\"" pod="openshift-multus/multus-4lmlt" podUID="9e857d8a-7289-4352-9de9-1b5d0bd21e8f" Feb 22 00:08:31 crc kubenswrapper[4857]: I0222 00:08:31.076779 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:31 crc kubenswrapper[4857]: E0222 00:08:31.077212 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:31 crc kubenswrapper[4857]: I0222 00:08:31.077277 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:31 crc kubenswrapper[4857]: I0222 00:08:31.077387 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:31 crc kubenswrapper[4857]: E0222 00:08:31.077734 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:31 crc kubenswrapper[4857]: E0222 00:08:31.077879 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:31 crc kubenswrapper[4857]: I0222 00:08:31.078103 4857 scope.go:117] "RemoveContainer" containerID="f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715" Feb 22 00:08:31 crc kubenswrapper[4857]: E0222 00:08:31.078272 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-6rc8w_openshift-ovn-kubernetes(9a0db241-7fc5-4d71-a8cc-534f4d303883)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" Feb 22 00:08:31 crc kubenswrapper[4857]: I0222 00:08:31.682144 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4lmlt_9e857d8a-7289-4352-9de9-1b5d0bd21e8f/kube-multus/1.log" Feb 22 00:08:32 crc kubenswrapper[4857]: I0222 00:08:32.077250 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:32 crc kubenswrapper[4857]: E0222 00:08:32.077388 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:33 crc kubenswrapper[4857]: E0222 00:08:33.066975 4857 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 22 00:08:33 crc kubenswrapper[4857]: I0222 00:08:33.077557 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:33 crc kubenswrapper[4857]: I0222 00:08:33.077604 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:33 crc kubenswrapper[4857]: I0222 00:08:33.077640 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:33 crc kubenswrapper[4857]: E0222 00:08:33.079370 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:33 crc kubenswrapper[4857]: E0222 00:08:33.079572 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:33 crc kubenswrapper[4857]: E0222 00:08:33.079801 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:33 crc kubenswrapper[4857]: E0222 00:08:33.159897 4857 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 22 00:08:34 crc kubenswrapper[4857]: I0222 00:08:34.076616 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:34 crc kubenswrapper[4857]: E0222 00:08:34.076861 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:35 crc kubenswrapper[4857]: I0222 00:08:35.077290 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:35 crc kubenswrapper[4857]: I0222 00:08:35.077307 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:35 crc kubenswrapper[4857]: E0222 00:08:35.077468 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:35 crc kubenswrapper[4857]: I0222 00:08:35.077514 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:35 crc kubenswrapper[4857]: E0222 00:08:35.077959 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:35 crc kubenswrapper[4857]: E0222 00:08:35.078400 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:36 crc kubenswrapper[4857]: I0222 00:08:36.076545 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:36 crc kubenswrapper[4857]: E0222 00:08:36.076725 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:37 crc kubenswrapper[4857]: I0222 00:08:37.077511 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:37 crc kubenswrapper[4857]: I0222 00:08:37.077614 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:37 crc kubenswrapper[4857]: E0222 00:08:37.077649 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:37 crc kubenswrapper[4857]: I0222 00:08:37.077725 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:37 crc kubenswrapper[4857]: E0222 00:08:37.077808 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:37 crc kubenswrapper[4857]: E0222 00:08:37.077866 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:38 crc kubenswrapper[4857]: I0222 00:08:38.077132 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:38 crc kubenswrapper[4857]: E0222 00:08:38.077316 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:38 crc kubenswrapper[4857]: E0222 00:08:38.161646 4857 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 22 00:08:39 crc kubenswrapper[4857]: I0222 00:08:39.076989 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:39 crc kubenswrapper[4857]: I0222 00:08:39.077129 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:39 crc kubenswrapper[4857]: E0222 00:08:39.077193 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:39 crc kubenswrapper[4857]: E0222 00:08:39.077295 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:39 crc kubenswrapper[4857]: I0222 00:08:39.077339 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:39 crc kubenswrapper[4857]: E0222 00:08:39.077538 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:40 crc kubenswrapper[4857]: I0222 00:08:40.076837 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:40 crc kubenswrapper[4857]: E0222 00:08:40.077029 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:41 crc kubenswrapper[4857]: I0222 00:08:41.077540 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:41 crc kubenswrapper[4857]: E0222 00:08:41.077934 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:41 crc kubenswrapper[4857]: I0222 00:08:41.077713 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:41 crc kubenswrapper[4857]: I0222 00:08:41.077646 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:41 crc kubenswrapper[4857]: E0222 00:08:41.078001 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:41 crc kubenswrapper[4857]: E0222 00:08:41.078210 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:42 crc kubenswrapper[4857]: I0222 00:08:42.076488 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:42 crc kubenswrapper[4857]: E0222 00:08:42.076824 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:43 crc kubenswrapper[4857]: I0222 00:08:43.077129 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:43 crc kubenswrapper[4857]: I0222 00:08:43.077196 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:43 crc kubenswrapper[4857]: I0222 00:08:43.078069 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:43 crc kubenswrapper[4857]: E0222 00:08:43.078212 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:43 crc kubenswrapper[4857]: E0222 00:08:43.078293 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:43 crc kubenswrapper[4857]: E0222 00:08:43.078408 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:43 crc kubenswrapper[4857]: E0222 00:08:43.162288 4857 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 22 00:08:44 crc kubenswrapper[4857]: I0222 00:08:44.076968 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:44 crc kubenswrapper[4857]: I0222 00:08:44.077335 4857 scope.go:117] "RemoveContainer" containerID="28685fd8939b71352e0be13ea9b6f38671b481eb13c5d08c79e788e8fc1c1f1c" Feb 22 00:08:44 crc kubenswrapper[4857]: E0222 00:08:44.077950 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:44 crc kubenswrapper[4857]: I0222 00:08:44.078805 4857 scope.go:117] "RemoveContainer" containerID="f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715" Feb 22 00:08:44 crc kubenswrapper[4857]: I0222 00:08:44.727209 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4lmlt_9e857d8a-7289-4352-9de9-1b5d0bd21e8f/kube-multus/1.log" Feb 22 00:08:44 crc kubenswrapper[4857]: I0222 00:08:44.727307 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4lmlt" event={"ID":"9e857d8a-7289-4352-9de9-1b5d0bd21e8f","Type":"ContainerStarted","Data":"57944544b90601bf139fb88d074d1c7fc35b6050e563aa238c6f1f02184473a4"} Feb 22 00:08:44 crc kubenswrapper[4857]: I0222 00:08:44.729819 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovnkube-controller/3.log" Feb 22 00:08:44 crc kubenswrapper[4857]: I0222 00:08:44.732778 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerStarted","Data":"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366"} Feb 22 00:08:44 crc kubenswrapper[4857]: I0222 00:08:44.733332 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:08:44 crc kubenswrapper[4857]: I0222 00:08:44.772001 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podStartSLOduration=108.77197954 podStartE2EDuration="1m48.77197954s" podCreationTimestamp="2026-02-22 00:06:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:44.768905082 +0000 UTC m=+132.407634345" watchObservedRunningTime="2026-02-22 00:08:44.77197954 +0000 UTC m=+132.410708803" Feb 22 00:08:44 crc kubenswrapper[4857]: I0222 00:08:44.908716 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-jnm7q"] Feb 22 00:08:44 crc kubenswrapper[4857]: I0222 00:08:44.908846 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:44 crc kubenswrapper[4857]: E0222 00:08:44.908947 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:45 crc kubenswrapper[4857]: I0222 00:08:45.077187 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:45 crc kubenswrapper[4857]: E0222 00:08:45.077527 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:45 crc kubenswrapper[4857]: I0222 00:08:45.077288 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:45 crc kubenswrapper[4857]: E0222 00:08:45.077590 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:46 crc kubenswrapper[4857]: I0222 00:08:46.077082 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:46 crc kubenswrapper[4857]: I0222 00:08:46.077116 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:46 crc kubenswrapper[4857]: E0222 00:08:46.077192 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:46 crc kubenswrapper[4857]: E0222 00:08:46.077387 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:47 crc kubenswrapper[4857]: I0222 00:08:47.076912 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:47 crc kubenswrapper[4857]: E0222 00:08:47.077198 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 22 00:08:47 crc kubenswrapper[4857]: I0222 00:08:47.077298 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:47 crc kubenswrapper[4857]: E0222 00:08:47.077424 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 22 00:08:48 crc kubenswrapper[4857]: I0222 00:08:48.077530 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:48 crc kubenswrapper[4857]: E0222 00:08:48.077832 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 22 00:08:48 crc kubenswrapper[4857]: I0222 00:08:48.078006 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:48 crc kubenswrapper[4857]: E0222 00:08:48.078445 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jnm7q" podUID="a0fb0fe8-ad7a-48a5-8a76-737cd5076f53" Feb 22 00:08:49 crc kubenswrapper[4857]: I0222 00:08:49.077454 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:08:49 crc kubenswrapper[4857]: I0222 00:08:49.077464 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:08:49 crc kubenswrapper[4857]: I0222 00:08:49.079393 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 22 00:08:49 crc kubenswrapper[4857]: I0222 00:08:49.080851 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 22 00:08:50 crc kubenswrapper[4857]: I0222 00:08:50.076511 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:08:50 crc kubenswrapper[4857]: I0222 00:08:50.076556 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:08:50 crc kubenswrapper[4857]: I0222 00:08:50.079613 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 22 00:08:50 crc kubenswrapper[4857]: I0222 00:08:50.079646 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 22 00:08:50 crc kubenswrapper[4857]: I0222 00:08:50.079851 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 22 00:08:50 crc kubenswrapper[4857]: I0222 00:08:50.080017 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.066544 4857 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.110319 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29528640-g7sjv"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.111074 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29528640-g7sjv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.116012 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-cwlf7"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.116667 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.117061 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.117638 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-cwlf7" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.118031 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.118901 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.119378 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.122292 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.122305 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.122369 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.122533 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.122766 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.123258 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk9zf"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.123602 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.124021 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk9zf" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.124222 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.124240 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.124495 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6vrqd"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.127130 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8sts2"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.128840 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.130446 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6vrqd" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.131290 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.133882 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-h4pmt"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.137754 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-mx94d"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.158683 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-h4pmt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.160094 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-c669b"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.160424 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.160628 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nj6dw"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.160669 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.161436 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nj6dw" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.161821 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.162050 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.162114 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.162232 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.162320 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.162419 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.162561 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.162663 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.162799 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.163005 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.163021 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.163178 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.164069 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.164184 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.164466 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.165132 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.165788 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.165949 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-ctjrv"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.166466 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.168771 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.168843 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.172714 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.172884 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.173142 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.173324 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.173401 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.173594 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.173681 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.173790 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.173802 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.174724 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.176824 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.176931 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.177079 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.177150 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.176836 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.177222 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.177161 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.176836 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.177383 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.177555 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.180451 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.181001 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4nvsx"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.181010 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.181279 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6d9wr"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.181626 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6d9wr" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.181938 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.182143 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4nvsx" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.183190 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.183310 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.183395 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.184275 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.184376 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.184455 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.184538 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.184707 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.184794 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.184881 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.184960 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.185064 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.185153 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.185225 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.185956 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.186179 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.186271 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.186502 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.187164 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.187411 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.187476 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.187536 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.188617 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.188744 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.188904 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.189467 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.189723 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.190309 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-wq8b8"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.190897 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.191144 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.192509 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.192622 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.192718 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.192840 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g4v7w"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.193364 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g4v7w" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.205463 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.206440 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.207596 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.208910 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.209137 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.209331 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.209609 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.209882 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.211192 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-96hns"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.212807 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.215837 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.216828 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.217023 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.217481 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.217482 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.231260 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.232178 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.232353 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.232478 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.234016 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.234212 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.234733 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.237727 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.237837 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.237995 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6xx4f"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.238637 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6xx4f" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.243597 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.249950 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.250508 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-wc47x"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.250956 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2p58w"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.251539 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2p58w" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.252138 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.252315 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-wc47x" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.254200 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vmk2g"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.254706 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.256482 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8x7sr"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.257098 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-8x7sr" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.258628 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.259057 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.260599 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lm22m"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.261005 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.265339 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.265748 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.267058 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-99cq6"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.267339 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4drbd"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.267664 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4drbd" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.267812 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-99cq6" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.268876 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.269139 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.269309 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.278915 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.279816 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-m6mzr"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.280244 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-m6mzr" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.280399 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.284230 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-lfd8v"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.284757 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.288839 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.290316 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.291002 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.291659 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5a31ed34-05d3-4e0f-bb4b-03d086e01f84-etcd-ca\") pod \"etcd-operator-b45778765-ctjrv\" (UID: \"5a31ed34-05d3-4e0f-bb4b-03d086e01f84\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.291697 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4tjq\" (UniqueName: \"kubernetes.io/projected/5fb885c7-2a02-4adb-af59-35d8d9a1a3d9-kube-api-access-z4tjq\") pod \"openshift-controller-manager-operator-756b6f6bc6-6vrqd\" (UID: \"5fb885c7-2a02-4adb-af59-35d8d9a1a3d9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6vrqd" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.291724 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60d12108-1ec8-4074-a60c-2b143a83f59f-config\") pod \"route-controller-manager-6576b87f9c-p979v\" (UID: \"60d12108-1ec8-4074-a60c-2b143a83f59f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.291746 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60d12108-1ec8-4074-a60c-2b143a83f59f-serving-cert\") pod \"route-controller-manager-6576b87f9c-p979v\" (UID: \"60d12108-1ec8-4074-a60c-2b143a83f59f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.291767 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chd2q\" (UniqueName: \"kubernetes.io/projected/860d5a18-25c6-4484-bc45-2e741cd6b708-kube-api-access-chd2q\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.291790 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fef422b7-19df-431d-b574-73457050f022-service-ca-bundle\") pod \"authentication-operator-69f744f599-8sts2\" (UID: \"fef422b7-19df-431d-b574-73457050f022\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.291811 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/860d5a18-25c6-4484-bc45-2e741cd6b708-node-pullsecrets\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.291832 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/860d5a18-25c6-4484-bc45-2e741cd6b708-audit\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.291856 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rwf8\" (UniqueName: \"kubernetes.io/projected/e5ba4e2d-a81f-4ab0-be7c-6ae18ba8f3cf-kube-api-access-6rwf8\") pod \"cluster-samples-operator-665b6dd947-zk9zf\" (UID: \"e5ba4e2d-a81f-4ab0-be7c-6ae18ba8f3cf\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk9zf" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.291878 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c719ea96-d4bf-4f14-b8bd-c32bb6a20c46-auth-proxy-config\") pod \"machine-approver-56656f9798-l9f6g\" (UID: \"c719ea96-d4bf-4f14-b8bd-c32bb6a20c46\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.291900 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-audit-policies\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.291920 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/860d5a18-25c6-4484-bc45-2e741cd6b708-image-import-ca\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.291938 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/860d5a18-25c6-4484-bc45-2e741cd6b708-serving-cert\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.291971 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fef422b7-19df-431d-b574-73457050f022-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8sts2\" (UID: \"fef422b7-19df-431d-b574-73457050f022\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.291992 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f5acab23-9994-4c18-bc50-4f5ec7c7f525-console-oauth-config\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292013 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f5acab23-9994-4c18-bc50-4f5ec7c7f525-trusted-ca-bundle\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292050 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5a31ed34-05d3-4e0f-bb4b-03d086e01f84-etcd-service-ca\") pod \"etcd-operator-b45778765-ctjrv\" (UID: \"5a31ed34-05d3-4e0f-bb4b-03d086e01f84\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292071 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/614b8dc2-9b2a-4425-b0d0-78292729e035-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-g4v7w\" (UID: \"614b8dc2-9b2a-4425-b0d0-78292729e035\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g4v7w" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292091 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e00acf3c-29bc-447d-bb80-7a1a8bf7cd84-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-4nvsx\" (UID: \"e00acf3c-29bc-447d-bb80-7a1a8bf7cd84\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4nvsx" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292108 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e8326fb9-1a3f-402a-acf2-62f7d6e82411-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-c669b\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292122 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-serving-cert\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292136 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnw74\" (UniqueName: \"kubernetes.io/projected/29721fdf-b040-414a-a6ca-6ba9dc544526-kube-api-access-dnw74\") pod \"console-operator-58897d9998-h4pmt\" (UID: \"29721fdf-b040-414a-a6ca-6ba9dc544526\") " pod="openshift-console-operator/console-operator-58897d9998-h4pmt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292152 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbwq7\" (UniqueName: \"kubernetes.io/projected/c719ea96-d4bf-4f14-b8bd-c32bb6a20c46-kube-api-access-tbwq7\") pod \"machine-approver-56656f9798-l9f6g\" (UID: \"c719ea96-d4bf-4f14-b8bd-c32bb6a20c46\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292167 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e5ba4e2d-a81f-4ab0-be7c-6ae18ba8f3cf-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-zk9zf\" (UID: \"e5ba4e2d-a81f-4ab0-be7c-6ae18ba8f3cf\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk9zf" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292185 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292205 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s48vw\" (UniqueName: \"kubernetes.io/projected/f07c5f6c-743d-4032-b70b-ecc0f514295d-kube-api-access-s48vw\") pod \"image-pruner-29528640-g7sjv\" (UID: \"f07c5f6c-743d-4032-b70b-ecc0f514295d\") " pod="openshift-image-registry/image-pruner-29528640-g7sjv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292223 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvbzl\" (UniqueName: \"kubernetes.io/projected/e8326fb9-1a3f-402a-acf2-62f7d6e82411-kube-api-access-jvbzl\") pod \"controller-manager-879f6c89f-c669b\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292238 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/614b8dc2-9b2a-4425-b0d0-78292729e035-config\") pod \"kube-controller-manager-operator-78b949d7b-g4v7w\" (UID: \"614b8dc2-9b2a-4425-b0d0-78292729e035\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g4v7w" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292254 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c719ea96-d4bf-4f14-b8bd-c32bb6a20c46-config\") pod \"machine-approver-56656f9798-l9f6g\" (UID: \"c719ea96-d4bf-4f14-b8bd-c32bb6a20c46\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292270 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/860d5a18-25c6-4484-bc45-2e741cd6b708-audit-dir\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292286 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9bce83c-ae40-482f-88d2-7de36a1a7ea5-config\") pod \"machine-api-operator-5694c8668f-cwlf7\" (UID: \"f9bce83c-ae40-482f-88d2-7de36a1a7ea5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cwlf7" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292301 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-audit-dir\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292316 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e8326fb9-1a3f-402a-acf2-62f7d6e82411-client-ca\") pod \"controller-manager-879f6c89f-c669b\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292332 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/706cd825-7128-4833-8260-b3b73cccaac7-trusted-ca\") pod \"ingress-operator-5b745b69d9-spwkv\" (UID: \"706cd825-7128-4833-8260-b3b73cccaac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292347 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fb885c7-2a02-4adb-af59-35d8d9a1a3d9-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6vrqd\" (UID: \"5fb885c7-2a02-4adb-af59-35d8d9a1a3d9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6vrqd" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292362 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-etcd-client\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292378 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29721fdf-b040-414a-a6ca-6ba9dc544526-serving-cert\") pod \"console-operator-58897d9998-h4pmt\" (UID: \"29721fdf-b040-414a-a6ca-6ba9dc544526\") " pod="openshift-console-operator/console-operator-58897d9998-h4pmt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292394 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5a31ed34-05d3-4e0f-bb4b-03d086e01f84-etcd-client\") pod \"etcd-operator-b45778765-ctjrv\" (UID: \"5a31ed34-05d3-4e0f-bb4b-03d086e01f84\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292409 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e00acf3c-29bc-447d-bb80-7a1a8bf7cd84-config\") pod \"kube-apiserver-operator-766d6c64bb-4nvsx\" (UID: \"e00acf3c-29bc-447d-bb80-7a1a8bf7cd84\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4nvsx" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292425 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f9bce83c-ae40-482f-88d2-7de36a1a7ea5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-cwlf7\" (UID: \"f9bce83c-ae40-482f-88d2-7de36a1a7ea5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cwlf7" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292443 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/4edfa330-6a4a-4f86-b606-db0adb8301c8-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4dw8f\" (UID: \"4edfa330-6a4a-4f86-b606-db0adb8301c8\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292467 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdfvv\" (UniqueName: \"kubernetes.io/projected/c2b40552-4007-4981-ba86-2d8864fc20cf-kube-api-access-xdfvv\") pod \"openshift-apiserver-operator-796bbdcf4f-nj6dw\" (UID: \"c2b40552-4007-4981-ba86-2d8864fc20cf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nj6dw" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292482 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-encryption-config\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292497 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29721fdf-b040-414a-a6ca-6ba9dc544526-trusted-ca\") pod \"console-operator-58897d9998-h4pmt\" (UID: \"29721fdf-b040-414a-a6ca-6ba9dc544526\") " pod="openshift-console-operator/console-operator-58897d9998-h4pmt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292515 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f5acab23-9994-4c18-bc50-4f5ec7c7f525-console-config\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292530 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29721fdf-b040-414a-a6ca-6ba9dc544526-config\") pod \"console-operator-58897d9998-h4pmt\" (UID: \"29721fdf-b040-414a-a6ca-6ba9dc544526\") " pod="openshift-console-operator/console-operator-58897d9998-h4pmt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292546 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/860d5a18-25c6-4484-bc45-2e741cd6b708-encryption-config\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292560 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/706cd825-7128-4833-8260-b3b73cccaac7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-spwkv\" (UID: \"706cd825-7128-4833-8260-b3b73cccaac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292579 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fa7f27c-4fa5-45db-b376-99c5c6468ec2-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6d9wr\" (UID: \"0fa7f27c-4fa5-45db-b376-99c5c6468ec2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6d9wr" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292594 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/614b8dc2-9b2a-4425-b0d0-78292729e035-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-g4v7w\" (UID: \"614b8dc2-9b2a-4425-b0d0-78292729e035\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g4v7w" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292610 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-769sl\" (UniqueName: \"kubernetes.io/projected/fef422b7-19df-431d-b574-73457050f022-kube-api-access-769sl\") pod \"authentication-operator-69f744f599-8sts2\" (UID: \"fef422b7-19df-431d-b574-73457050f022\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292627 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c719ea96-d4bf-4f14-b8bd-c32bb6a20c46-machine-approver-tls\") pod \"machine-approver-56656f9798-l9f6g\" (UID: \"c719ea96-d4bf-4f14-b8bd-c32bb6a20c46\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292641 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4edfa330-6a4a-4f86-b606-db0adb8301c8-serving-cert\") pod \"openshift-config-operator-7777fb866f-4dw8f\" (UID: \"4edfa330-6a4a-4f86-b606-db0adb8301c8\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292656 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f5acab23-9994-4c18-bc50-4f5ec7c7f525-oauth-serving-cert\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292672 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fef422b7-19df-431d-b574-73457050f022-serving-cert\") pod \"authentication-operator-69f744f599-8sts2\" (UID: \"fef422b7-19df-431d-b574-73457050f022\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292694 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8326fb9-1a3f-402a-acf2-62f7d6e82411-serving-cert\") pod \"controller-manager-879f6c89f-c669b\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292708 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a31ed34-05d3-4e0f-bb4b-03d086e01f84-config\") pod \"etcd-operator-b45778765-ctjrv\" (UID: \"5a31ed34-05d3-4e0f-bb4b-03d086e01f84\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292724 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292738 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8326fb9-1a3f-402a-acf2-62f7d6e82411-config\") pod \"controller-manager-879f6c89f-c669b\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292753 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2b40552-4007-4981-ba86-2d8864fc20cf-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-nj6dw\" (UID: \"c2b40552-4007-4981-ba86-2d8864fc20cf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nj6dw" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292768 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/860d5a18-25c6-4484-bc45-2e741cd6b708-etcd-serving-ca\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292785 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0fa7f27c-4fa5-45db-b376-99c5c6468ec2-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6d9wr\" (UID: \"0fa7f27c-4fa5-45db-b376-99c5c6468ec2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6d9wr" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292802 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2b40552-4007-4981-ba86-2d8864fc20cf-config\") pod \"openshift-apiserver-operator-796bbdcf4f-nj6dw\" (UID: \"c2b40552-4007-4981-ba86-2d8864fc20cf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nj6dw" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292817 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f07c5f6c-743d-4032-b70b-ecc0f514295d-serviceca\") pod \"image-pruner-29528640-g7sjv\" (UID: \"f07c5f6c-743d-4032-b70b-ecc0f514295d\") " pod="openshift-image-registry/image-pruner-29528640-g7sjv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292832 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf7md\" (UniqueName: \"kubernetes.io/projected/f9bce83c-ae40-482f-88d2-7de36a1a7ea5-kube-api-access-rf7md\") pod \"machine-api-operator-5694c8668f-cwlf7\" (UID: \"f9bce83c-ae40-482f-88d2-7de36a1a7ea5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cwlf7" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292853 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60d12108-1ec8-4074-a60c-2b143a83f59f-client-ca\") pod \"route-controller-manager-6576b87f9c-p979v\" (UID: \"60d12108-1ec8-4074-a60c-2b143a83f59f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292868 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/860d5a18-25c6-4484-bc45-2e741cd6b708-config\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292882 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f9bce83c-ae40-482f-88d2-7de36a1a7ea5-images\") pod \"machine-api-operator-5694c8668f-cwlf7\" (UID: \"f9bce83c-ae40-482f-88d2-7de36a1a7ea5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cwlf7" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292896 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a31ed34-05d3-4e0f-bb4b-03d086e01f84-serving-cert\") pod \"etcd-operator-b45778765-ctjrv\" (UID: \"5a31ed34-05d3-4e0f-bb4b-03d086e01f84\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292912 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tths\" (UniqueName: \"kubernetes.io/projected/60d12108-1ec8-4074-a60c-2b143a83f59f-kube-api-access-7tths\") pod \"route-controller-manager-6576b87f9c-p979v\" (UID: \"60d12108-1ec8-4074-a60c-2b143a83f59f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292926 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f5acab23-9994-4c18-bc50-4f5ec7c7f525-service-ca\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292941 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fef422b7-19df-431d-b574-73457050f022-config\") pod \"authentication-operator-69f744f599-8sts2\" (UID: \"fef422b7-19df-431d-b574-73457050f022\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292967 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/860d5a18-25c6-4484-bc45-2e741cd6b708-trusted-ca-bundle\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.292989 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/706cd825-7128-4833-8260-b3b73cccaac7-metrics-tls\") pod \"ingress-operator-5b745b69d9-spwkv\" (UID: \"706cd825-7128-4833-8260-b3b73cccaac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.293003 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fb885c7-2a02-4adb-af59-35d8d9a1a3d9-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6vrqd\" (UID: \"5fb885c7-2a02-4adb-af59-35d8d9a1a3d9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6vrqd" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.293022 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wncn\" (UniqueName: \"kubernetes.io/projected/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-kube-api-access-5wncn\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.293050 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/860d5a18-25c6-4484-bc45-2e741cd6b708-etcd-client\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.293066 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fa7f27c-4fa5-45db-b376-99c5c6468ec2-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6d9wr\" (UID: \"0fa7f27c-4fa5-45db-b376-99c5c6468ec2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6d9wr" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.293080 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6wcn\" (UniqueName: \"kubernetes.io/projected/4edfa330-6a4a-4f86-b606-db0adb8301c8-kube-api-access-z6wcn\") pod \"openshift-config-operator-7777fb866f-4dw8f\" (UID: \"4edfa330-6a4a-4f86-b606-db0adb8301c8\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.293099 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c74dt\" (UniqueName: \"kubernetes.io/projected/5a31ed34-05d3-4e0f-bb4b-03d086e01f84-kube-api-access-c74dt\") pod \"etcd-operator-b45778765-ctjrv\" (UID: \"5a31ed34-05d3-4e0f-bb4b-03d086e01f84\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.293113 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djdj9\" (UniqueName: \"kubernetes.io/projected/706cd825-7128-4833-8260-b3b73cccaac7-kube-api-access-djdj9\") pod \"ingress-operator-5b745b69d9-spwkv\" (UID: \"706cd825-7128-4833-8260-b3b73cccaac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.293129 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e00acf3c-29bc-447d-bb80-7a1a8bf7cd84-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-4nvsx\" (UID: \"e00acf3c-29bc-447d-bb80-7a1a8bf7cd84\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4nvsx" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.293145 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f5acab23-9994-4c18-bc50-4f5ec7c7f525-console-serving-cert\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.293163 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8whq\" (UniqueName: \"kubernetes.io/projected/f5acab23-9994-4c18-bc50-4f5ec7c7f525-kube-api-access-h8whq\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.296408 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-d4kmf"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.296908 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4kmf" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.297545 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.298417 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.298505 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.301258 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-tctk6"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.302696 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tctk6" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.309297 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-h8pq2"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.310331 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-h8pq2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.313462 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-cwlf7"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.327029 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.327090 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6vrqd"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.328949 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.337529 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk9zf"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.341782 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-c4xt9"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.344267 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.345760 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-c4xt9" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.346348 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-p9ztw"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.348077 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-p9ztw" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.349777 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-h4pmt"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.357110 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.358613 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.358801 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4nvsx"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.360050 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-c669b"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.362294 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6xx4f"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.362754 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g4v7w"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.364139 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.365465 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-99cq6"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.366927 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.368500 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2p58w"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.371376 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8sts2"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.372575 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.373156 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nj6dw"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.374161 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-ctjrv"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.375745 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wq8b8"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.376685 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.377237 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.378356 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-96hns"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.379387 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.380397 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.381391 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6d9wr"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.382481 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vmk2g"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.383625 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lm22m"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.385513 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4drbd"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.386978 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-mx94d"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.388242 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8x7sr"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.389419 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-f6z92"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.390248 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-f6z92" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.390651 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-m6mzr"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.391834 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9x24j"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.393509 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29528640-g7sjv"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.393615 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.393881 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/860d5a18-25c6-4484-bc45-2e741cd6b708-trusted-ca-bundle\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.393923 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026-srv-cert\") pod \"catalog-operator-68c6474976-nqshz\" (UID: \"2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.393948 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87g9m\" (UniqueName: \"kubernetes.io/projected/ff701efb-cabc-479a-ac05-4e9334e70385-kube-api-access-87g9m\") pod \"downloads-7954f5f757-h8pq2\" (UID: \"ff701efb-cabc-479a-ac05-4e9334e70385\") " pod="openshift-console/downloads-7954f5f757-h8pq2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.393955 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.393974 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6ab42c49-5c7e-4bfb-9a4b-53e6eab21a33-metrics-tls\") pod \"dns-operator-744455d44c-wc47x\" (UID: \"6ab42c49-5c7e-4bfb-9a4b-53e6eab21a33\") " pod="openshift-dns-operator/dns-operator-744455d44c-wc47x" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.393993 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56twm\" (UniqueName: \"kubernetes.io/projected/e8865dea-cf78-4acd-bcb4-aad234172e99-kube-api-access-56twm\") pod \"migrator-59844c95c7-tctk6\" (UID: \"e8865dea-cf78-4acd-bcb4-aad234172e99\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tctk6" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394031 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/706cd825-7128-4833-8260-b3b73cccaac7-metrics-tls\") pod \"ingress-operator-5b745b69d9-spwkv\" (UID: \"706cd825-7128-4833-8260-b3b73cccaac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394062 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fb885c7-2a02-4adb-af59-35d8d9a1a3d9-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6vrqd\" (UID: \"5fb885c7-2a02-4adb-af59-35d8d9a1a3d9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6vrqd" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394079 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c5c791b9-d379-4311-b961-0927ca48702a-srv-cert\") pod \"olm-operator-6b444d44fb-7wd8c\" (UID: \"c5c791b9-d379-4311-b961-0927ca48702a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394102 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wncn\" (UniqueName: \"kubernetes.io/projected/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-kube-api-access-5wncn\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394117 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/860d5a18-25c6-4484-bc45-2e741cd6b708-etcd-client\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394135 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ft7b\" (UniqueName: \"kubernetes.io/projected/9e531e33-264a-498e-b279-cf687f1e6327-kube-api-access-4ft7b\") pod \"service-ca-9c57cc56f-99cq6\" (UID: \"9e531e33-264a-498e-b279-cf687f1e6327\") " pod="openshift-service-ca/service-ca-9c57cc56f-99cq6" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394152 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57hqs\" (UniqueName: \"kubernetes.io/projected/05ee3f6d-2b50-4da4-8688-01ae8364d166-kube-api-access-57hqs\") pod \"collect-profiles-29528640-sbzs9\" (UID: \"05ee3f6d-2b50-4da4-8688-01ae8364d166\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394171 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c74dt\" (UniqueName: \"kubernetes.io/projected/5a31ed34-05d3-4e0f-bb4b-03d086e01f84-kube-api-access-c74dt\") pod \"etcd-operator-b45778765-ctjrv\" (UID: \"5a31ed34-05d3-4e0f-bb4b-03d086e01f84\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394186 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djdj9\" (UniqueName: \"kubernetes.io/projected/706cd825-7128-4833-8260-b3b73cccaac7-kube-api-access-djdj9\") pod \"ingress-operator-5b745b69d9-spwkv\" (UID: \"706cd825-7128-4833-8260-b3b73cccaac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394204 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fa7f27c-4fa5-45db-b376-99c5c6468ec2-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6d9wr\" (UID: \"0fa7f27c-4fa5-45db-b376-99c5c6468ec2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6d9wr" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394222 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6wcn\" (UniqueName: \"kubernetes.io/projected/4edfa330-6a4a-4f86-b606-db0adb8301c8-kube-api-access-z6wcn\") pod \"openshift-config-operator-7777fb866f-4dw8f\" (UID: \"4edfa330-6a4a-4f86-b606-db0adb8301c8\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394240 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026-profile-collector-cert\") pod \"catalog-operator-68c6474976-nqshz\" (UID: \"2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394258 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05ee3f6d-2b50-4da4-8688-01ae8364d166-secret-volume\") pod \"collect-profiles-29528640-sbzs9\" (UID: \"05ee3f6d-2b50-4da4-8688-01ae8364d166\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394277 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f5acab23-9994-4c18-bc50-4f5ec7c7f525-console-serving-cert\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394294 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8whq\" (UniqueName: \"kubernetes.io/projected/f5acab23-9994-4c18-bc50-4f5ec7c7f525-kube-api-access-h8whq\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394309 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e00acf3c-29bc-447d-bb80-7a1a8bf7cd84-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-4nvsx\" (UID: \"e00acf3c-29bc-447d-bb80-7a1a8bf7cd84\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4nvsx" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394327 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5a31ed34-05d3-4e0f-bb4b-03d086e01f84-etcd-ca\") pod \"etcd-operator-b45778765-ctjrv\" (UID: \"5a31ed34-05d3-4e0f-bb4b-03d086e01f84\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394346 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/13d16a24-fc30-4d19-a3d5-1c55cc0021e4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8x7sr\" (UID: \"13d16a24-fc30-4d19-a3d5-1c55cc0021e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8x7sr" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394365 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4tjq\" (UniqueName: \"kubernetes.io/projected/5fb885c7-2a02-4adb-af59-35d8d9a1a3d9-kube-api-access-z4tjq\") pod \"openshift-controller-manager-operator-756b6f6bc6-6vrqd\" (UID: \"5fb885c7-2a02-4adb-af59-35d8d9a1a3d9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6vrqd" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394384 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60d12108-1ec8-4074-a60c-2b143a83f59f-config\") pod \"route-controller-manager-6576b87f9c-p979v\" (UID: \"60d12108-1ec8-4074-a60c-2b143a83f59f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394403 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60d12108-1ec8-4074-a60c-2b143a83f59f-serving-cert\") pod \"route-controller-manager-6576b87f9c-p979v\" (UID: \"60d12108-1ec8-4074-a60c-2b143a83f59f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394421 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chd2q\" (UniqueName: \"kubernetes.io/projected/860d5a18-25c6-4484-bc45-2e741cd6b708-kube-api-access-chd2q\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394449 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fef422b7-19df-431d-b574-73457050f022-service-ca-bundle\") pod \"authentication-operator-69f744f599-8sts2\" (UID: \"fef422b7-19df-431d-b574-73457050f022\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394468 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/860d5a18-25c6-4484-bc45-2e741cd6b708-node-pullsecrets\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394562 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/860d5a18-25c6-4484-bc45-2e741cd6b708-audit\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394584 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrrjf\" (UniqueName: \"kubernetes.io/projected/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-kube-api-access-wrrjf\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394603 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c719ea96-d4bf-4f14-b8bd-c32bb6a20c46-auth-proxy-config\") pod \"machine-approver-56656f9798-l9f6g\" (UID: \"c719ea96-d4bf-4f14-b8bd-c32bb6a20c46\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394635 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rwf8\" (UniqueName: \"kubernetes.io/projected/e5ba4e2d-a81f-4ab0-be7c-6ae18ba8f3cf-kube-api-access-6rwf8\") pod \"cluster-samples-operator-665b6dd947-zk9zf\" (UID: \"e5ba4e2d-a81f-4ab0-be7c-6ae18ba8f3cf\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk9zf" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394663 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fef422b7-19df-431d-b574-73457050f022-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8sts2\" (UID: \"fef422b7-19df-431d-b574-73457050f022\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394680 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-audit-policies\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394698 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/860d5a18-25c6-4484-bc45-2e741cd6b708-image-import-ca\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394712 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/860d5a18-25c6-4484-bc45-2e741cd6b708-serving-cert\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394730 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqbd5\" (UniqueName: \"kubernetes.io/projected/61fba77d-c313-4c67-a0d8-7fbf2e172aa4-kube-api-access-lqbd5\") pod \"package-server-manager-789f6589d5-2p58w\" (UID: \"61fba77d-c313-4c67-a0d8-7fbf2e172aa4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2p58w" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394747 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e00acf3c-29bc-447d-bb80-7a1a8bf7cd84-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-4nvsx\" (UID: \"e00acf3c-29bc-447d-bb80-7a1a8bf7cd84\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4nvsx" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394764 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e8326fb9-1a3f-402a-acf2-62f7d6e82411-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-c669b\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394782 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f5acab23-9994-4c18-bc50-4f5ec7c7f525-console-oauth-config\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394797 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f5acab23-9994-4c18-bc50-4f5ec7c7f525-trusted-ca-bundle\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394813 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5a31ed34-05d3-4e0f-bb4b-03d086e01f84-etcd-service-ca\") pod \"etcd-operator-b45778765-ctjrv\" (UID: \"5a31ed34-05d3-4e0f-bb4b-03d086e01f84\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394829 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/614b8dc2-9b2a-4425-b0d0-78292729e035-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-g4v7w\" (UID: \"614b8dc2-9b2a-4425-b0d0-78292729e035\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g4v7w" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394845 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-serving-cert\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394861 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnw74\" (UniqueName: \"kubernetes.io/projected/29721fdf-b040-414a-a6ca-6ba9dc544526-kube-api-access-dnw74\") pod \"console-operator-58897d9998-h4pmt\" (UID: \"29721fdf-b040-414a-a6ca-6ba9dc544526\") " pod="openshift-console-operator/console-operator-58897d9998-h4pmt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394878 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/9e531e33-264a-498e-b279-cf687f1e6327-signing-key\") pod \"service-ca-9c57cc56f-99cq6\" (UID: \"9e531e33-264a-498e-b279-cf687f1e6327\") " pod="openshift-service-ca/service-ca-9c57cc56f-99cq6" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394896 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbwq7\" (UniqueName: \"kubernetes.io/projected/c719ea96-d4bf-4f14-b8bd-c32bb6a20c46-kube-api-access-tbwq7\") pod \"machine-approver-56656f9798-l9f6g\" (UID: \"c719ea96-d4bf-4f14-b8bd-c32bb6a20c46\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394912 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e5ba4e2d-a81f-4ab0-be7c-6ae18ba8f3cf-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-zk9zf\" (UID: \"e5ba4e2d-a81f-4ab0-be7c-6ae18ba8f3cf\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk9zf" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394927 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394943 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s48vw\" (UniqueName: \"kubernetes.io/projected/f07c5f6c-743d-4032-b70b-ecc0f514295d-kube-api-access-s48vw\") pod \"image-pruner-29528640-g7sjv\" (UID: \"f07c5f6c-743d-4032-b70b-ecc0f514295d\") " pod="openshift-image-registry/image-pruner-29528640-g7sjv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394958 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dqz9\" (UniqueName: \"kubernetes.io/projected/13d16a24-fc30-4d19-a3d5-1c55cc0021e4-kube-api-access-7dqz9\") pod \"multus-admission-controller-857f4d67dd-8x7sr\" (UID: \"13d16a24-fc30-4d19-a3d5-1c55cc0021e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8x7sr" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394976 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvbzl\" (UniqueName: \"kubernetes.io/projected/e8326fb9-1a3f-402a-acf2-62f7d6e82411-kube-api-access-jvbzl\") pod \"controller-manager-879f6c89f-c669b\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.394993 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/614b8dc2-9b2a-4425-b0d0-78292729e035-config\") pod \"kube-controller-manager-operator-78b949d7b-g4v7w\" (UID: \"614b8dc2-9b2a-4425-b0d0-78292729e035\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g4v7w" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395009 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c719ea96-d4bf-4f14-b8bd-c32bb6a20c46-config\") pod \"machine-approver-56656f9798-l9f6g\" (UID: \"c719ea96-d4bf-4f14-b8bd-c32bb6a20c46\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395025 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-audit-dir\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395058 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/860d5a18-25c6-4484-bc45-2e741cd6b708-audit-dir\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395074 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9bce83c-ae40-482f-88d2-7de36a1a7ea5-config\") pod \"machine-api-operator-5694c8668f-cwlf7\" (UID: \"f9bce83c-ae40-482f-88d2-7de36a1a7ea5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cwlf7" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395088 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-etcd-client\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395107 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e8326fb9-1a3f-402a-acf2-62f7d6e82411-client-ca\") pod \"controller-manager-879f6c89f-c669b\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395122 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/706cd825-7128-4833-8260-b3b73cccaac7-trusted-ca\") pod \"ingress-operator-5b745b69d9-spwkv\" (UID: \"706cd825-7128-4833-8260-b3b73cccaac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395137 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fb885c7-2a02-4adb-af59-35d8d9a1a3d9-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6vrqd\" (UID: \"5fb885c7-2a02-4adb-af59-35d8d9a1a3d9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6vrqd" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395154 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29721fdf-b040-414a-a6ca-6ba9dc544526-serving-cert\") pod \"console-operator-58897d9998-h4pmt\" (UID: \"29721fdf-b040-414a-a6ca-6ba9dc544526\") " pod="openshift-console-operator/console-operator-58897d9998-h4pmt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395170 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgqrl\" (UniqueName: \"kubernetes.io/projected/2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026-kube-api-access-pgqrl\") pod \"catalog-operator-68c6474976-nqshz\" (UID: \"2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395188 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f9bce83c-ae40-482f-88d2-7de36a1a7ea5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-cwlf7\" (UID: \"f9bce83c-ae40-482f-88d2-7de36a1a7ea5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cwlf7" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395206 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5a31ed34-05d3-4e0f-bb4b-03d086e01f84-etcd-client\") pod \"etcd-operator-b45778765-ctjrv\" (UID: \"5a31ed34-05d3-4e0f-bb4b-03d086e01f84\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395221 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e00acf3c-29bc-447d-bb80-7a1a8bf7cd84-config\") pod \"kube-apiserver-operator-766d6c64bb-4nvsx\" (UID: \"e00acf3c-29bc-447d-bb80-7a1a8bf7cd84\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4nvsx" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395266 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e6ba7763-9bd5-45dd-a9a9-d47dc7f381ba-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-4drbd\" (UID: \"e6ba7763-9bd5-45dd-a9a9-d47dc7f381ba\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4drbd" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395293 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/4edfa330-6a4a-4f86-b606-db0adb8301c8-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4dw8f\" (UID: \"4edfa330-6a4a-4f86-b606-db0adb8301c8\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395309 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8thm\" (UniqueName: \"kubernetes.io/projected/c5c791b9-d379-4311-b961-0927ca48702a-kube-api-access-f8thm\") pod \"olm-operator-6b444d44fb-7wd8c\" (UID: \"c5c791b9-d379-4311-b961-0927ca48702a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395325 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c5c791b9-d379-4311-b961-0927ca48702a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-7wd8c\" (UID: \"c5c791b9-d379-4311-b961-0927ca48702a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395341 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f5acab23-9994-4c18-bc50-4f5ec7c7f525-console-config\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395369 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdfvv\" (UniqueName: \"kubernetes.io/projected/c2b40552-4007-4981-ba86-2d8864fc20cf-kube-api-access-xdfvv\") pod \"openshift-apiserver-operator-796bbdcf4f-nj6dw\" (UID: \"c2b40552-4007-4981-ba86-2d8864fc20cf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nj6dw" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395385 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-encryption-config\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395400 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29721fdf-b040-414a-a6ca-6ba9dc544526-trusted-ca\") pod \"console-operator-58897d9998-h4pmt\" (UID: \"29721fdf-b040-414a-a6ca-6ba9dc544526\") " pod="openshift-console-operator/console-operator-58897d9998-h4pmt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395417 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xn9z\" (UniqueName: \"kubernetes.io/projected/6ab42c49-5c7e-4bfb-9a4b-53e6eab21a33-kube-api-access-6xn9z\") pod \"dns-operator-744455d44c-wc47x\" (UID: \"6ab42c49-5c7e-4bfb-9a4b-53e6eab21a33\") " pod="openshift-dns-operator/dns-operator-744455d44c-wc47x" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395441 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/706cd825-7128-4833-8260-b3b73cccaac7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-spwkv\" (UID: \"706cd825-7128-4833-8260-b3b73cccaac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395496 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fa7f27c-4fa5-45db-b376-99c5c6468ec2-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6d9wr\" (UID: \"0fa7f27c-4fa5-45db-b376-99c5c6468ec2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6d9wr" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395530 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29721fdf-b040-414a-a6ca-6ba9dc544526-config\") pod \"console-operator-58897d9998-h4pmt\" (UID: \"29721fdf-b040-414a-a6ca-6ba9dc544526\") " pod="openshift-console-operator/console-operator-58897d9998-h4pmt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395546 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/860d5a18-25c6-4484-bc45-2e741cd6b708-encryption-config\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395563 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-769sl\" (UniqueName: \"kubernetes.io/projected/fef422b7-19df-431d-b574-73457050f022-kube-api-access-769sl\") pod \"authentication-operator-69f744f599-8sts2\" (UID: \"fef422b7-19df-431d-b574-73457050f022\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395580 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/614b8dc2-9b2a-4425-b0d0-78292729e035-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-g4v7w\" (UID: \"614b8dc2-9b2a-4425-b0d0-78292729e035\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g4v7w" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395598 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c719ea96-d4bf-4f14-b8bd-c32bb6a20c46-machine-approver-tls\") pod \"machine-approver-56656f9798-l9f6g\" (UID: \"c719ea96-d4bf-4f14-b8bd-c32bb6a20c46\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395615 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4edfa330-6a4a-4f86-b606-db0adb8301c8-serving-cert\") pod \"openshift-config-operator-7777fb866f-4dw8f\" (UID: \"4edfa330-6a4a-4f86-b606-db0adb8301c8\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395633 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f5acab23-9994-4c18-bc50-4f5ec7c7f525-oauth-serving-cert\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395650 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fef422b7-19df-431d-b574-73457050f022-serving-cert\") pod \"authentication-operator-69f744f599-8sts2\" (UID: \"fef422b7-19df-431d-b574-73457050f022\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395668 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/9e531e33-264a-498e-b279-cf687f1e6327-signing-cabundle\") pod \"service-ca-9c57cc56f-99cq6\" (UID: \"9e531e33-264a-498e-b279-cf687f1e6327\") " pod="openshift-service-ca/service-ca-9c57cc56f-99cq6" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395693 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-default-certificate\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395696 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/860d5a18-25c6-4484-bc45-2e741cd6b708-trusted-ca-bundle\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395712 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a31ed34-05d3-4e0f-bb4b-03d086e01f84-config\") pod \"etcd-operator-b45778765-ctjrv\" (UID: \"5a31ed34-05d3-4e0f-bb4b-03d086e01f84\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395758 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8326fb9-1a3f-402a-acf2-62f7d6e82411-serving-cert\") pod \"controller-manager-879f6c89f-c669b\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395787 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395811 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-service-ca-bundle\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395835 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8326fb9-1a3f-402a-acf2-62f7d6e82411-config\") pod \"controller-manager-879f6c89f-c669b\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395862 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2b40552-4007-4981-ba86-2d8864fc20cf-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-nj6dw\" (UID: \"c2b40552-4007-4981-ba86-2d8864fc20cf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nj6dw" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395884 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/860d5a18-25c6-4484-bc45-2e741cd6b708-etcd-serving-ca\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395902 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0fa7f27c-4fa5-45db-b376-99c5c6468ec2-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6d9wr\" (UID: \"0fa7f27c-4fa5-45db-b376-99c5c6468ec2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6d9wr" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395932 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2b40552-4007-4981-ba86-2d8864fc20cf-config\") pod \"openshift-apiserver-operator-796bbdcf4f-nj6dw\" (UID: \"c2b40552-4007-4981-ba86-2d8864fc20cf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nj6dw" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395956 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f07c5f6c-743d-4032-b70b-ecc0f514295d-serviceca\") pod \"image-pruner-29528640-g7sjv\" (UID: \"f07c5f6c-743d-4032-b70b-ecc0f514295d\") " pod="openshift-image-registry/image-pruner-29528640-g7sjv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.395980 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf7md\" (UniqueName: \"kubernetes.io/projected/f9bce83c-ae40-482f-88d2-7de36a1a7ea5-kube-api-access-rf7md\") pod \"machine-api-operator-5694c8668f-cwlf7\" (UID: \"f9bce83c-ae40-482f-88d2-7de36a1a7ea5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cwlf7" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.396005 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-stats-auth\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.396023 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60d12108-1ec8-4074-a60c-2b143a83f59f-client-ca\") pod \"route-controller-manager-6576b87f9c-p979v\" (UID: \"60d12108-1ec8-4074-a60c-2b143a83f59f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.396072 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/860d5a18-25c6-4484-bc45-2e741cd6b708-config\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.396093 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f9bce83c-ae40-482f-88d2-7de36a1a7ea5-images\") pod \"machine-api-operator-5694c8668f-cwlf7\" (UID: \"f9bce83c-ae40-482f-88d2-7de36a1a7ea5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cwlf7" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.396109 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a31ed34-05d3-4e0f-bb4b-03d086e01f84-serving-cert\") pod \"etcd-operator-b45778765-ctjrv\" (UID: \"5a31ed34-05d3-4e0f-bb4b-03d086e01f84\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.396124 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tths\" (UniqueName: \"kubernetes.io/projected/60d12108-1ec8-4074-a60c-2b143a83f59f-kube-api-access-7tths\") pod \"route-controller-manager-6576b87f9c-p979v\" (UID: \"60d12108-1ec8-4074-a60c-2b143a83f59f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.396141 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/61fba77d-c313-4c67-a0d8-7fbf2e172aa4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2p58w\" (UID: \"61fba77d-c313-4c67-a0d8-7fbf2e172aa4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2p58w" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.396161 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f5acab23-9994-4c18-bc50-4f5ec7c7f525-service-ca\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.396176 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fef422b7-19df-431d-b574-73457050f022-config\") pod \"authentication-operator-69f744f599-8sts2\" (UID: \"fef422b7-19df-431d-b574-73457050f022\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.396195 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwrvg\" (UniqueName: \"kubernetes.io/projected/e6ba7763-9bd5-45dd-a9a9-d47dc7f381ba-kube-api-access-qwrvg\") pod \"control-plane-machine-set-operator-78cbb6b69f-4drbd\" (UID: \"e6ba7763-9bd5-45dd-a9a9-d47dc7f381ba\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4drbd" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.396212 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05ee3f6d-2b50-4da4-8688-01ae8364d166-config-volume\") pod \"collect-profiles-29528640-sbzs9\" (UID: \"05ee3f6d-2b50-4da4-8688-01ae8364d166\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.396403 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-metrics-certs\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.396520 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a31ed34-05d3-4e0f-bb4b-03d086e01f84-config\") pod \"etcd-operator-b45778765-ctjrv\" (UID: \"5a31ed34-05d3-4e0f-bb4b-03d086e01f84\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.397089 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fb885c7-2a02-4adb-af59-35d8d9a1a3d9-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6vrqd\" (UID: \"5fb885c7-2a02-4adb-af59-35d8d9a1a3d9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6vrqd" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.397790 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-tctk6"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.397813 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-c4xt9"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.398192 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/4edfa330-6a4a-4f86-b606-db0adb8301c8-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4dw8f\" (UID: \"4edfa330-6a4a-4f86-b606-db0adb8301c8\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.398840 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29721fdf-b040-414a-a6ca-6ba9dc544526-config\") pod \"console-operator-58897d9998-h4pmt\" (UID: \"29721fdf-b040-414a-a6ca-6ba9dc544526\") " pod="openshift-console-operator/console-operator-58897d9998-h4pmt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.399555 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.400216 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e8326fb9-1a3f-402a-acf2-62f7d6e82411-client-ca\") pod \"controller-manager-879f6c89f-c669b\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.400309 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fa7f27c-4fa5-45db-b376-99c5c6468ec2-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6d9wr\" (UID: \"0fa7f27c-4fa5-45db-b376-99c5c6468ec2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6d9wr" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.400357 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.401303 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e5ba4e2d-a81f-4ab0-be7c-6ae18ba8f3cf-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-zk9zf\" (UID: \"e5ba4e2d-a81f-4ab0-be7c-6ae18ba8f3cf\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk9zf" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.401358 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f5acab23-9994-4c18-bc50-4f5ec7c7f525-console-config\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.401379 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8326fb9-1a3f-402a-acf2-62f7d6e82411-config\") pod \"controller-manager-879f6c89f-c669b\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.401539 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60d12108-1ec8-4074-a60c-2b143a83f59f-client-ca\") pod \"route-controller-manager-6576b87f9c-p979v\" (UID: \"60d12108-1ec8-4074-a60c-2b143a83f59f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.402021 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/860d5a18-25c6-4484-bc45-2e741cd6b708-config\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.402649 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f9bce83c-ae40-482f-88d2-7de36a1a7ea5-images\") pod \"machine-api-operator-5694c8668f-cwlf7\" (UID: \"f9bce83c-ae40-482f-88d2-7de36a1a7ea5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cwlf7" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.403299 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/706cd825-7128-4833-8260-b3b73cccaac7-trusted-ca\") pod \"ingress-operator-5b745b69d9-spwkv\" (UID: \"706cd825-7128-4833-8260-b3b73cccaac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.403529 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-etcd-client\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.404225 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-encryption-config\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.404251 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c719ea96-d4bf-4f14-b8bd-c32bb6a20c46-config\") pod \"machine-approver-56656f9798-l9f6g\" (UID: \"c719ea96-d4bf-4f14-b8bd-c32bb6a20c46\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.404281 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-audit-dir\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.404304 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/860d5a18-25c6-4484-bc45-2e741cd6b708-audit-dir\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.404579 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fa7f27c-4fa5-45db-b376-99c5c6468ec2-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6d9wr\" (UID: \"0fa7f27c-4fa5-45db-b376-99c5c6468ec2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6d9wr" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.405073 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9bce83c-ae40-482f-88d2-7de36a1a7ea5-config\") pod \"machine-api-operator-5694c8668f-cwlf7\" (UID: \"f9bce83c-ae40-482f-88d2-7de36a1a7ea5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cwlf7" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.405502 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4edfa330-6a4a-4f86-b606-db0adb8301c8-serving-cert\") pod \"openshift-config-operator-7777fb866f-4dw8f\" (UID: \"4edfa330-6a4a-4f86-b606-db0adb8301c8\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.406474 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.406550 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a31ed34-05d3-4e0f-bb4b-03d086e01f84-serving-cert\") pod \"etcd-operator-b45778765-ctjrv\" (UID: \"5a31ed34-05d3-4e0f-bb4b-03d086e01f84\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.406661 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8326fb9-1a3f-402a-acf2-62f7d6e82411-serving-cert\") pod \"controller-manager-879f6c89f-c669b\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.407347 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/29721fdf-b040-414a-a6ca-6ba9dc544526-trusted-ca\") pod \"console-operator-58897d9998-h4pmt\" (UID: \"29721fdf-b040-414a-a6ca-6ba9dc544526\") " pod="openshift-console-operator/console-operator-58897d9998-h4pmt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.407396 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f5acab23-9994-4c18-bc50-4f5ec7c7f525-oauth-serving-cert\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.407657 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fef422b7-19df-431d-b574-73457050f022-service-ca-bundle\") pod \"authentication-operator-69f744f599-8sts2\" (UID: \"fef422b7-19df-431d-b574-73457050f022\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.408549 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f5acab23-9994-4c18-bc50-4f5ec7c7f525-console-serving-cert\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.408621 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/706cd825-7128-4833-8260-b3b73cccaac7-metrics-tls\") pod \"ingress-operator-5b745b69d9-spwkv\" (UID: \"706cd825-7128-4833-8260-b3b73cccaac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.409107 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fef422b7-19df-431d-b574-73457050f022-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-8sts2\" (UID: \"fef422b7-19df-431d-b574-73457050f022\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.409506 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f9bce83c-ae40-482f-88d2-7de36a1a7ea5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-cwlf7\" (UID: \"f9bce83c-ae40-482f-88d2-7de36a1a7ea5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cwlf7" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.409669 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/860d5a18-25c6-4484-bc45-2e741cd6b708-etcd-serving-ca\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.410400 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60d12108-1ec8-4074-a60c-2b143a83f59f-config\") pod \"route-controller-manager-6576b87f9c-p979v\" (UID: \"60d12108-1ec8-4074-a60c-2b143a83f59f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.410494 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e00acf3c-29bc-447d-bb80-7a1a8bf7cd84-config\") pod \"kube-apiserver-operator-766d6c64bb-4nvsx\" (UID: \"e00acf3c-29bc-447d-bb80-7a1a8bf7cd84\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4nvsx" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.410918 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e00acf3c-29bc-447d-bb80-7a1a8bf7cd84-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-4nvsx\" (UID: \"e00acf3c-29bc-447d-bb80-7a1a8bf7cd84\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4nvsx" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.410947 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29721fdf-b040-414a-a6ca-6ba9dc544526-serving-cert\") pod \"console-operator-58897d9998-h4pmt\" (UID: \"29721fdf-b040-414a-a6ca-6ba9dc544526\") " pod="openshift-console-operator/console-operator-58897d9998-h4pmt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.411140 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f5acab23-9994-4c18-bc50-4f5ec7c7f525-trusted-ca-bundle\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.411216 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/860d5a18-25c6-4484-bc45-2e741cd6b708-node-pullsecrets\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.411427 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5a31ed34-05d3-4e0f-bb4b-03d086e01f84-etcd-ca\") pod \"etcd-operator-b45778765-ctjrv\" (UID: \"5a31ed34-05d3-4e0f-bb4b-03d086e01f84\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.411916 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2b40552-4007-4981-ba86-2d8864fc20cf-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-nj6dw\" (UID: \"c2b40552-4007-4981-ba86-2d8864fc20cf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nj6dw" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.412029 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/860d5a18-25c6-4484-bc45-2e741cd6b708-encryption-config\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.412572 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-audit-policies\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.412628 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f07c5f6c-743d-4032-b70b-ecc0f514295d-serviceca\") pod \"image-pruner-29528640-g7sjv\" (UID: \"f07c5f6c-743d-4032-b70b-ecc0f514295d\") " pod="openshift-image-registry/image-pruner-29528640-g7sjv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.412876 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2b40552-4007-4981-ba86-2d8864fc20cf-config\") pod \"openshift-apiserver-operator-796bbdcf4f-nj6dw\" (UID: \"c2b40552-4007-4981-ba86-2d8864fc20cf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nj6dw" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.413145 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/860d5a18-25c6-4484-bc45-2e741cd6b708-audit\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.413590 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5a31ed34-05d3-4e0f-bb4b-03d086e01f84-etcd-client\") pod \"etcd-operator-b45778765-ctjrv\" (UID: \"5a31ed34-05d3-4e0f-bb4b-03d086e01f84\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.413775 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fb885c7-2a02-4adb-af59-35d8d9a1a3d9-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6vrqd\" (UID: \"5fb885c7-2a02-4adb-af59-35d8d9a1a3d9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6vrqd" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.413950 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c719ea96-d4bf-4f14-b8bd-c32bb6a20c46-machine-approver-tls\") pod \"machine-approver-56656f9798-l9f6g\" (UID: \"c719ea96-d4bf-4f14-b8bd-c32bb6a20c46\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.414931 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c719ea96-d4bf-4f14-b8bd-c32bb6a20c46-auth-proxy-config\") pod \"machine-approver-56656f9798-l9f6g\" (UID: \"c719ea96-d4bf-4f14-b8bd-c32bb6a20c46\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.414978 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f5acab23-9994-4c18-bc50-4f5ec7c7f525-service-ca\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.415083 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e8326fb9-1a3f-402a-acf2-62f7d6e82411-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-c669b\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.415429 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5a31ed34-05d3-4e0f-bb4b-03d086e01f84-etcd-service-ca\") pod \"etcd-operator-b45778765-ctjrv\" (UID: \"5a31ed34-05d3-4e0f-bb4b-03d086e01f84\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.415678 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/860d5a18-25c6-4484-bc45-2e741cd6b708-serving-cert\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.416236 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/860d5a18-25c6-4484-bc45-2e741cd6b708-image-import-ca\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.416442 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.416571 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fef422b7-19df-431d-b574-73457050f022-config\") pod \"authentication-operator-69f744f599-8sts2\" (UID: \"fef422b7-19df-431d-b574-73457050f022\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.416682 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fef422b7-19df-431d-b574-73457050f022-serving-cert\") pod \"authentication-operator-69f744f599-8sts2\" (UID: \"fef422b7-19df-431d-b574-73457050f022\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.417212 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-p9ztw"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.417664 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60d12108-1ec8-4074-a60c-2b143a83f59f-serving-cert\") pod \"route-controller-manager-6576b87f9c-p979v\" (UID: \"60d12108-1ec8-4074-a60c-2b143a83f59f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.419223 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f5acab23-9994-4c18-bc50-4f5ec7c7f525-console-oauth-config\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.419264 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-h8pq2"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.421131 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.421932 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/614b8dc2-9b2a-4425-b0d0-78292729e035-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-g4v7w\" (UID: \"614b8dc2-9b2a-4425-b0d0-78292729e035\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g4v7w" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.422780 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-d4kmf"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.424414 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-wc47x"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.424428 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/860d5a18-25c6-4484-bc45-2e741cd6b708-etcd-client\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.424450 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-serving-cert\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.426426 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9x24j"] Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.434373 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.444789 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/614b8dc2-9b2a-4425-b0d0-78292729e035-config\") pod \"kube-controller-manager-operator-78b949d7b-g4v7w\" (UID: \"614b8dc2-9b2a-4425-b0d0-78292729e035\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g4v7w" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.467210 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.473397 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.496904 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/9e531e33-264a-498e-b279-cf687f1e6327-signing-cabundle\") pod \"service-ca-9c57cc56f-99cq6\" (UID: \"9e531e33-264a-498e-b279-cf687f1e6327\") " pod="openshift-service-ca/service-ca-9c57cc56f-99cq6" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.496935 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-default-certificate\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.496962 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-service-ca-bundle\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.496990 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-stats-auth\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497019 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwrvg\" (UniqueName: \"kubernetes.io/projected/e6ba7763-9bd5-45dd-a9a9-d47dc7f381ba-kube-api-access-qwrvg\") pod \"control-plane-machine-set-operator-78cbb6b69f-4drbd\" (UID: \"e6ba7763-9bd5-45dd-a9a9-d47dc7f381ba\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4drbd" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497052 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/61fba77d-c313-4c67-a0d8-7fbf2e172aa4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2p58w\" (UID: \"61fba77d-c313-4c67-a0d8-7fbf2e172aa4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2p58w" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497070 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05ee3f6d-2b50-4da4-8688-01ae8364d166-config-volume\") pod \"collect-profiles-29528640-sbzs9\" (UID: \"05ee3f6d-2b50-4da4-8688-01ae8364d166\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497084 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-metrics-certs\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497101 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87g9m\" (UniqueName: \"kubernetes.io/projected/ff701efb-cabc-479a-ac05-4e9334e70385-kube-api-access-87g9m\") pod \"downloads-7954f5f757-h8pq2\" (UID: \"ff701efb-cabc-479a-ac05-4e9334e70385\") " pod="openshift-console/downloads-7954f5f757-h8pq2" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497117 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026-srv-cert\") pod \"catalog-operator-68c6474976-nqshz\" (UID: \"2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497133 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6ab42c49-5c7e-4bfb-9a4b-53e6eab21a33-metrics-tls\") pod \"dns-operator-744455d44c-wc47x\" (UID: \"6ab42c49-5c7e-4bfb-9a4b-53e6eab21a33\") " pod="openshift-dns-operator/dns-operator-744455d44c-wc47x" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497153 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56twm\" (UniqueName: \"kubernetes.io/projected/e8865dea-cf78-4acd-bcb4-aad234172e99-kube-api-access-56twm\") pod \"migrator-59844c95c7-tctk6\" (UID: \"e8865dea-cf78-4acd-bcb4-aad234172e99\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tctk6" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497179 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c5c791b9-d379-4311-b961-0927ca48702a-srv-cert\") pod \"olm-operator-6b444d44fb-7wd8c\" (UID: \"c5c791b9-d379-4311-b961-0927ca48702a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497199 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57hqs\" (UniqueName: \"kubernetes.io/projected/05ee3f6d-2b50-4da4-8688-01ae8364d166-kube-api-access-57hqs\") pod \"collect-profiles-29528640-sbzs9\" (UID: \"05ee3f6d-2b50-4da4-8688-01ae8364d166\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497222 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ft7b\" (UniqueName: \"kubernetes.io/projected/9e531e33-264a-498e-b279-cf687f1e6327-kube-api-access-4ft7b\") pod \"service-ca-9c57cc56f-99cq6\" (UID: \"9e531e33-264a-498e-b279-cf687f1e6327\") " pod="openshift-service-ca/service-ca-9c57cc56f-99cq6" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497417 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026-profile-collector-cert\") pod \"catalog-operator-68c6474976-nqshz\" (UID: \"2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497431 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05ee3f6d-2b50-4da4-8688-01ae8364d166-secret-volume\") pod \"collect-profiles-29528640-sbzs9\" (UID: \"05ee3f6d-2b50-4da4-8688-01ae8364d166\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497474 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/13d16a24-fc30-4d19-a3d5-1c55cc0021e4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8x7sr\" (UID: \"13d16a24-fc30-4d19-a3d5-1c55cc0021e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8x7sr" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497493 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrrjf\" (UniqueName: \"kubernetes.io/projected/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-kube-api-access-wrrjf\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497519 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqbd5\" (UniqueName: \"kubernetes.io/projected/61fba77d-c313-4c67-a0d8-7fbf2e172aa4-kube-api-access-lqbd5\") pod \"package-server-manager-789f6589d5-2p58w\" (UID: \"61fba77d-c313-4c67-a0d8-7fbf2e172aa4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2p58w" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497577 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/9e531e33-264a-498e-b279-cf687f1e6327-signing-key\") pod \"service-ca-9c57cc56f-99cq6\" (UID: \"9e531e33-264a-498e-b279-cf687f1e6327\") " pod="openshift-service-ca/service-ca-9c57cc56f-99cq6" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497594 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dqz9\" (UniqueName: \"kubernetes.io/projected/13d16a24-fc30-4d19-a3d5-1c55cc0021e4-kube-api-access-7dqz9\") pod \"multus-admission-controller-857f4d67dd-8x7sr\" (UID: \"13d16a24-fc30-4d19-a3d5-1c55cc0021e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8x7sr" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497625 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgqrl\" (UniqueName: \"kubernetes.io/projected/2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026-kube-api-access-pgqrl\") pod \"catalog-operator-68c6474976-nqshz\" (UID: \"2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497673 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e6ba7763-9bd5-45dd-a9a9-d47dc7f381ba-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-4drbd\" (UID: \"e6ba7763-9bd5-45dd-a9a9-d47dc7f381ba\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4drbd" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497700 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8thm\" (UniqueName: \"kubernetes.io/projected/c5c791b9-d379-4311-b961-0927ca48702a-kube-api-access-f8thm\") pod \"olm-operator-6b444d44fb-7wd8c\" (UID: \"c5c791b9-d379-4311-b961-0927ca48702a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497734 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xn9z\" (UniqueName: \"kubernetes.io/projected/6ab42c49-5c7e-4bfb-9a4b-53e6eab21a33-kube-api-access-6xn9z\") pod \"dns-operator-744455d44c-wc47x\" (UID: \"6ab42c49-5c7e-4bfb-9a4b-53e6eab21a33\") " pod="openshift-dns-operator/dns-operator-744455d44c-wc47x" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.497751 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c5c791b9-d379-4311-b961-0927ca48702a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-7wd8c\" (UID: \"c5c791b9-d379-4311-b961-0927ca48702a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.515137 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.533102 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.553173 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.573987 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.593619 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.614173 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.633608 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.653642 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.675225 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.681607 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/61fba77d-c313-4c67-a0d8-7fbf2e172aa4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2p58w\" (UID: \"61fba77d-c313-4c67-a0d8-7fbf2e172aa4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2p58w" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.694311 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.714474 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.733717 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.753950 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.762108 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6ab42c49-5c7e-4bfb-9a4b-53e6eab21a33-metrics-tls\") pod \"dns-operator-744455d44c-wc47x\" (UID: \"6ab42c49-5c7e-4bfb-9a4b-53e6eab21a33\") " pod="openshift-dns-operator/dns-operator-744455d44c-wc47x" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.774786 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.793261 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.817902 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.834575 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.864652 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.874673 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.895797 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.901011 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/13d16a24-fc30-4d19-a3d5-1c55cc0021e4-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8x7sr\" (UID: \"13d16a24-fc30-4d19-a3d5-1c55cc0021e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8x7sr" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.915856 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.934708 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.953102 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.973931 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 22 00:08:52 crc kubenswrapper[4857]: I0222 00:08:52.993627 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.014522 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.034210 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.054214 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.079456 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.094704 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.128516 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.134128 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.154066 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.173989 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.194352 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.223148 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.233409 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.254716 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.262168 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c5c791b9-d379-4311-b961-0927ca48702a-srv-cert\") pod \"olm-operator-6b444d44fb-7wd8c\" (UID: \"c5c791b9-d379-4311-b961-0927ca48702a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.272031 4857 request.go:700] Waited for 1.006098564s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dpprof-cert&limit=500&resourceVersion=0 Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.273476 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.280726 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05ee3f6d-2b50-4da4-8688-01ae8364d166-secret-volume\") pod \"collect-profiles-29528640-sbzs9\" (UID: \"05ee3f6d-2b50-4da4-8688-01ae8364d166\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.281235 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c5c791b9-d379-4311-b961-0927ca48702a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-7wd8c\" (UID: \"c5c791b9-d379-4311-b961-0927ca48702a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.281276 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026-profile-collector-cert\") pod \"catalog-operator-68c6474976-nqshz\" (UID: \"2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.293843 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.300823 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e6ba7763-9bd5-45dd-a9a9-d47dc7f381ba-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-4drbd\" (UID: \"e6ba7763-9bd5-45dd-a9a9-d47dc7f381ba\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4drbd" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.313651 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.318403 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/9e531e33-264a-498e-b279-cf687f1e6327-signing-cabundle\") pod \"service-ca-9c57cc56f-99cq6\" (UID: \"9e531e33-264a-498e-b279-cf687f1e6327\") " pod="openshift-service-ca/service-ca-9c57cc56f-99cq6" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.334403 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.353669 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.374246 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.380791 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/9e531e33-264a-498e-b279-cf687f1e6327-signing-key\") pod \"service-ca-9c57cc56f-99cq6\" (UID: \"9e531e33-264a-498e-b279-cf687f1e6327\") " pod="openshift-service-ca/service-ca-9c57cc56f-99cq6" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.394071 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.415615 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.433930 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.454420 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.474487 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.493541 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 22 00:08:53 crc kubenswrapper[4857]: E0222 00:08:53.497285 4857 secret.go:188] Couldn't get secret openshift-ingress/router-stats-default: failed to sync secret cache: timed out waiting for the condition Feb 22 00:08:53 crc kubenswrapper[4857]: E0222 00:08:53.497343 4857 secret.go:188] Couldn't get secret openshift-ingress/router-metrics-certs-default: failed to sync secret cache: timed out waiting for the condition Feb 22 00:08:53 crc kubenswrapper[4857]: E0222 00:08:53.497371 4857 secret.go:188] Couldn't get secret openshift-ingress/router-certs-default: failed to sync secret cache: timed out waiting for the condition Feb 22 00:08:53 crc kubenswrapper[4857]: E0222 00:08:53.497427 4857 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Feb 22 00:08:53 crc kubenswrapper[4857]: E0222 00:08:53.497434 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-metrics-certs podName:1d0786e3-c21c-46f5-8ca9-7cad8ac62194 nodeName:}" failed. No retries permitted until 2026-02-22 00:08:53.997404158 +0000 UTC m=+141.636133451 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-metrics-certs") pod "router-default-5444994796-lfd8v" (UID: "1d0786e3-c21c-46f5-8ca9-7cad8ac62194") : failed to sync secret cache: timed out waiting for the condition Feb 22 00:08:53 crc kubenswrapper[4857]: E0222 00:08:53.497509 4857 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Feb 22 00:08:53 crc kubenswrapper[4857]: E0222 00:08:53.497524 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-stats-auth podName:1d0786e3-c21c-46f5-8ca9-7cad8ac62194 nodeName:}" failed. No retries permitted until 2026-02-22 00:08:53.997503991 +0000 UTC m=+141.636233234 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "stats-auth" (UniqueName: "kubernetes.io/secret/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-stats-auth") pod "router-default-5444994796-lfd8v" (UID: "1d0786e3-c21c-46f5-8ca9-7cad8ac62194") : failed to sync secret cache: timed out waiting for the condition Feb 22 00:08:53 crc kubenswrapper[4857]: E0222 00:08:53.497552 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026-srv-cert podName:2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026 nodeName:}" failed. No retries permitted until 2026-02-22 00:08:53.997546332 +0000 UTC m=+141.636275585 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026-srv-cert") pod "catalog-operator-68c6474976-nqshz" (UID: "2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026") : failed to sync secret cache: timed out waiting for the condition Feb 22 00:08:53 crc kubenswrapper[4857]: E0222 00:08:53.497517 4857 configmap.go:193] Couldn't get configMap openshift-ingress/service-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Feb 22 00:08:53 crc kubenswrapper[4857]: E0222 00:08:53.497571 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-default-certificate podName:1d0786e3-c21c-46f5-8ca9-7cad8ac62194 nodeName:}" failed. No retries permitted until 2026-02-22 00:08:53.997557952 +0000 UTC m=+141.636287215 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-certificate" (UniqueName: "kubernetes.io/secret/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-default-certificate") pod "router-default-5444994796-lfd8v" (UID: "1d0786e3-c21c-46f5-8ca9-7cad8ac62194") : failed to sync secret cache: timed out waiting for the condition Feb 22 00:08:53 crc kubenswrapper[4857]: E0222 00:08:53.497596 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/05ee3f6d-2b50-4da4-8688-01ae8364d166-config-volume podName:05ee3f6d-2b50-4da4-8688-01ae8364d166 nodeName:}" failed. No retries permitted until 2026-02-22 00:08:53.997586123 +0000 UTC m=+141.636315386 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/05ee3f6d-2b50-4da4-8688-01ae8364d166-config-volume") pod "collect-profiles-29528640-sbzs9" (UID: "05ee3f6d-2b50-4da4-8688-01ae8364d166") : failed to sync configmap cache: timed out waiting for the condition Feb 22 00:08:53 crc kubenswrapper[4857]: E0222 00:08:53.497611 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-service-ca-bundle podName:1d0786e3-c21c-46f5-8ca9-7cad8ac62194 nodeName:}" failed. No retries permitted until 2026-02-22 00:08:53.997604544 +0000 UTC m=+141.636333807 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-service-ca-bundle") pod "router-default-5444994796-lfd8v" (UID: "1d0786e3-c21c-46f5-8ca9-7cad8ac62194") : failed to sync configmap cache: timed out waiting for the condition Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.514185 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.533901 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.554098 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.573506 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.594794 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.613886 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.633488 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.654054 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.674676 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.694346 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.713883 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.733794 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.753685 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.774166 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.793938 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.814750 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.834169 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.854014 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.874750 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.893366 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.914204 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.933452 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.953972 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 22 00:08:53 crc kubenswrapper[4857]: I0222 00:08:53.993937 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.013634 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.019385 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-default-certificate\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.019451 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-service-ca-bundle\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.019518 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-stats-auth\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.019589 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05ee3f6d-2b50-4da4-8688-01ae8364d166-config-volume\") pod \"collect-profiles-29528640-sbzs9\" (UID: \"05ee3f6d-2b50-4da4-8688-01ae8364d166\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.019625 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-metrics-certs\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.019657 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026-srv-cert\") pod \"catalog-operator-68c6474976-nqshz\" (UID: \"2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.021162 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05ee3f6d-2b50-4da4-8688-01ae8364d166-config-volume\") pod \"collect-profiles-29528640-sbzs9\" (UID: \"05ee3f6d-2b50-4da4-8688-01ae8364d166\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.021764 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-service-ca-bundle\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.023577 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026-srv-cert\") pod \"catalog-operator-68c6474976-nqshz\" (UID: \"2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.024127 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-default-certificate\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.025743 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-metrics-certs\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.025970 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-stats-auth\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.035936 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.055031 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.074547 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.093719 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.113918 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.134264 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.154829 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.173634 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.194164 4857 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.214407 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.233546 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.271258 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wncn\" (UniqueName: \"kubernetes.io/projected/56f6f504-2f13-4ae2-b5cd-d67ec7c140a7-kube-api-access-5wncn\") pod \"apiserver-7bbb656c7d-9nx68\" (UID: \"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.271469 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.272101 4857 request.go:700] Waited for 1.873457684s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/serviceaccounts/route-controller-manager-sa/token Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.296371 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tths\" (UniqueName: \"kubernetes.io/projected/60d12108-1ec8-4074-a60c-2b143a83f59f-kube-api-access-7tths\") pod \"route-controller-manager-6576b87f9c-p979v\" (UID: \"60d12108-1ec8-4074-a60c-2b143a83f59f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.311538 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c74dt\" (UniqueName: \"kubernetes.io/projected/5a31ed34-05d3-4e0f-bb4b-03d086e01f84-kube-api-access-c74dt\") pod \"etcd-operator-b45778765-ctjrv\" (UID: \"5a31ed34-05d3-4e0f-bb4b-03d086e01f84\") " pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.330766 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djdj9\" (UniqueName: \"kubernetes.io/projected/706cd825-7128-4833-8260-b3b73cccaac7-kube-api-access-djdj9\") pod \"ingress-operator-5b745b69d9-spwkv\" (UID: \"706cd825-7128-4833-8260-b3b73cccaac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.355351 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdfvv\" (UniqueName: \"kubernetes.io/projected/c2b40552-4007-4981-ba86-2d8864fc20cf-kube-api-access-xdfvv\") pod \"openshift-apiserver-operator-796bbdcf4f-nj6dw\" (UID: \"c2b40552-4007-4981-ba86-2d8864fc20cf\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nj6dw" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.364349 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.370018 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s48vw\" (UniqueName: \"kubernetes.io/projected/f07c5f6c-743d-4032-b70b-ecc0f514295d-kube-api-access-s48vw\") pod \"image-pruner-29528640-g7sjv\" (UID: \"f07c5f6c-743d-4032-b70b-ecc0f514295d\") " pod="openshift-image-registry/image-pruner-29528640-g7sjv" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.392194 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvbzl\" (UniqueName: \"kubernetes.io/projected/e8326fb9-1a3f-402a-acf2-62f7d6e82411-kube-api-access-jvbzl\") pod \"controller-manager-879f6c89f-c669b\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.408757 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6wcn\" (UniqueName: \"kubernetes.io/projected/4edfa330-6a4a-4f86-b606-db0adb8301c8-kube-api-access-z6wcn\") pod \"openshift-config-operator-7777fb866f-4dw8f\" (UID: \"4edfa330-6a4a-4f86-b606-db0adb8301c8\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.429853 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8whq\" (UniqueName: \"kubernetes.io/projected/f5acab23-9994-4c18-bc50-4f5ec7c7f525-kube-api-access-h8whq\") pod \"console-f9d7485db-wq8b8\" (UID: \"f5acab23-9994-4c18-bc50-4f5ec7c7f525\") " pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.451650 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-769sl\" (UniqueName: \"kubernetes.io/projected/fef422b7-19df-431d-b574-73457050f022-kube-api-access-769sl\") pod \"authentication-operator-69f744f599-8sts2\" (UID: \"fef422b7-19df-431d-b574-73457050f022\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.458100 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.467559 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4tjq\" (UniqueName: \"kubernetes.io/projected/5fb885c7-2a02-4adb-af59-35d8d9a1a3d9-kube-api-access-z4tjq\") pod \"openshift-controller-manager-operator-756b6f6bc6-6vrqd\" (UID: \"5fb885c7-2a02-4adb-af59-35d8d9a1a3d9\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6vrqd" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.480368 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nj6dw" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.487421 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0fa7f27c-4fa5-45db-b376-99c5c6468ec2-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6d9wr\" (UID: \"0fa7f27c-4fa5-45db-b376-99c5c6468ec2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6d9wr" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.489293 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.495650 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.505985 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6d9wr" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.510951 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chd2q\" (UniqueName: \"kubernetes.io/projected/860d5a18-25c6-4484-bc45-2e741cd6b708-kube-api-access-chd2q\") pod \"apiserver-76f77b778f-mx94d\" (UID: \"860d5a18-25c6-4484-bc45-2e741cd6b708\") " pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.531879 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e00acf3c-29bc-447d-bb80-7a1a8bf7cd84-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-4nvsx\" (UID: \"e00acf3c-29bc-447d-bb80-7a1a8bf7cd84\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4nvsx" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.539503 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v"] Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.550240 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf7md\" (UniqueName: \"kubernetes.io/projected/f9bce83c-ae40-482f-88d2-7de36a1a7ea5-kube-api-access-rf7md\") pod \"machine-api-operator-5694c8668f-cwlf7\" (UID: \"f9bce83c-ae40-482f-88d2-7de36a1a7ea5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-cwlf7" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.564853 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4nvsx" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.569151 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbwq7\" (UniqueName: \"kubernetes.io/projected/c719ea96-d4bf-4f14-b8bd-c32bb6a20c46-kube-api-access-tbwq7\") pod \"machine-approver-56656f9798-l9f6g\" (UID: \"c719ea96-d4bf-4f14-b8bd-c32bb6a20c46\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.578752 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.594276 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/706cd825-7128-4833-8260-b3b73cccaac7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-spwkv\" (UID: \"706cd825-7128-4833-8260-b3b73cccaac7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.611323 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29528640-g7sjv" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.611644 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnw74\" (UniqueName: \"kubernetes.io/projected/29721fdf-b040-414a-a6ca-6ba9dc544526-kube-api-access-dnw74\") pod \"console-operator-58897d9998-h4pmt\" (UID: \"29721fdf-b040-414a-a6ca-6ba9dc544526\") " pod="openshift-console-operator/console-operator-58897d9998-h4pmt" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.632483 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rwf8\" (UniqueName: \"kubernetes.io/projected/e5ba4e2d-a81f-4ab0-be7c-6ae18ba8f3cf-kube-api-access-6rwf8\") pod \"cluster-samples-operator-665b6dd947-zk9zf\" (UID: \"e5ba4e2d-a81f-4ab0-be7c-6ae18ba8f3cf\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk9zf" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.656077 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-c669b"] Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.656575 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/614b8dc2-9b2a-4425-b0d0-78292729e035-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-g4v7w\" (UID: \"614b8dc2-9b2a-4425-b0d0-78292729e035\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g4v7w" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.664296 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-cwlf7" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.677208 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.687761 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68"] Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.694932 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk9zf" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.702883 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwrvg\" (UniqueName: \"kubernetes.io/projected/e6ba7763-9bd5-45dd-a9a9-d47dc7f381ba-kube-api-access-qwrvg\") pod \"control-plane-machine-set-operator-78cbb6b69f-4drbd\" (UID: \"e6ba7763-9bd5-45dd-a9a9-d47dc7f381ba\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4drbd" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.714285 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57hqs\" (UniqueName: \"kubernetes.io/projected/05ee3f6d-2b50-4da4-8688-01ae8364d166-kube-api-access-57hqs\") pod \"collect-profiles-29528640-sbzs9\" (UID: \"05ee3f6d-2b50-4da4-8688-01ae8364d166\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.716830 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.728973 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6vrqd" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.730339 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56twm\" (UniqueName: \"kubernetes.io/projected/e8865dea-cf78-4acd-bcb4-aad234172e99-kube-api-access-56twm\") pod \"migrator-59844c95c7-tctk6\" (UID: \"e8865dea-cf78-4acd-bcb4-aad234172e99\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tctk6" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.749160 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-h4pmt" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.768474 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87g9m\" (UniqueName: \"kubernetes.io/projected/ff701efb-cabc-479a-ac05-4e9334e70385-kube-api-access-87g9m\") pod \"downloads-7954f5f757-h8pq2\" (UID: \"ff701efb-cabc-479a-ac05-4e9334e70385\") " pod="openshift-console/downloads-7954f5f757-h8pq2" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.768897 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ft7b\" (UniqueName: \"kubernetes.io/projected/9e531e33-264a-498e-b279-cf687f1e6327-kube-api-access-4ft7b\") pod \"service-ca-9c57cc56f-99cq6\" (UID: \"9e531e33-264a-498e-b279-cf687f1e6327\") " pod="openshift-service-ca/service-ca-9c57cc56f-99cq6" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.771238 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.787778 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" event={"ID":"e8326fb9-1a3f-402a-acf2-62f7d6e82411","Type":"ContainerStarted","Data":"056c2a52ec7cc59cccff869b4363924e9e2d3cf007b48b239c60317d275e2467"} Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.791775 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrrjf\" (UniqueName: \"kubernetes.io/projected/1d0786e3-c21c-46f5-8ca9-7cad8ac62194-kube-api-access-wrrjf\") pod \"router-default-5444994796-lfd8v\" (UID: \"1d0786e3-c21c-46f5-8ca9-7cad8ac62194\") " pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.798194 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g" event={"ID":"c719ea96-d4bf-4f14-b8bd-c32bb6a20c46","Type":"ContainerStarted","Data":"d36daf74ad0ddd05b973695d5606a8140649cac90df5a18d6f12434fa58b0d07"} Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.806142 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" event={"ID":"60d12108-1ec8-4074-a60c-2b143a83f59f","Type":"ContainerStarted","Data":"49a036fb35334ac9f46de95f3f6cbedf36f22ab4de0f9a12eba85808c11618f0"} Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.815246 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" event={"ID":"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7","Type":"ContainerStarted","Data":"98f391818f4174155978d360f594eabe8a21b30c062642ff85451a2db802dd54"} Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.822173 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqbd5\" (UniqueName: \"kubernetes.io/projected/61fba77d-c313-4c67-a0d8-7fbf2e172aa4-kube-api-access-lqbd5\") pod \"package-server-manager-789f6589d5-2p58w\" (UID: \"61fba77d-c313-4c67-a0d8-7fbf2e172aa4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2p58w" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.841763 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dqz9\" (UniqueName: \"kubernetes.io/projected/13d16a24-fc30-4d19-a3d5-1c55cc0021e4-kube-api-access-7dqz9\") pod \"multus-admission-controller-857f4d67dd-8x7sr\" (UID: \"13d16a24-fc30-4d19-a3d5-1c55cc0021e4\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8x7sr" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.850281 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgqrl\" (UniqueName: \"kubernetes.io/projected/2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026-kube-api-access-pgqrl\") pod \"catalog-operator-68c6474976-nqshz\" (UID: \"2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.854678 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.869694 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8thm\" (UniqueName: \"kubernetes.io/projected/c5c791b9-d379-4311-b961-0927ca48702a-kube-api-access-f8thm\") pod \"olm-operator-6b444d44fb-7wd8c\" (UID: \"c5c791b9-d379-4311-b961-0927ca48702a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.884885 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g4v7w" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.890534 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xn9z\" (UniqueName: \"kubernetes.io/projected/6ab42c49-5c7e-4bfb-9a4b-53e6eab21a33-kube-api-access-6xn9z\") pod \"dns-operator-744455d44c-wc47x\" (UID: \"6ab42c49-5c7e-4bfb-9a4b-53e6eab21a33\") " pod="openshift-dns-operator/dns-operator-744455d44c-wc47x" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.937017 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nj6dw"] Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.938722 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f"] Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.939183 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2p58w" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.939301 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-wc47x" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.939318 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-8x7sr" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.941976 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mjr6\" (UniqueName: \"kubernetes.io/projected/9e0d9776-fb39-4996-9d7d-48a045cd2da9-kube-api-access-6mjr6\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.942554 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/46391957-feff-4548-8c20-c713712585af-proxy-tls\") pod \"machine-config-controller-84d6567774-6xx4f\" (UID: \"46391957-feff-4548-8c20-c713712585af\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6xx4f" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.942617 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9e0d9776-fb39-4996-9d7d-48a045cd2da9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.942670 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/46391957-feff-4548-8c20-c713712585af-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6xx4f\" (UID: \"46391957-feff-4548-8c20-c713712585af\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6xx4f" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.942691 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.942743 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5nh9\" (UniqueName: \"kubernetes.io/projected/46391957-feff-4548-8c20-c713712585af-kube-api-access-p5nh9\") pod \"machine-config-controller-84d6567774-6xx4f\" (UID: \"46391957-feff-4548-8c20-c713712585af\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6xx4f" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.942807 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9e0d9776-fb39-4996-9d7d-48a045cd2da9-registry-certificates\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.942843 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9e0d9776-fb39-4996-9d7d-48a045cd2da9-bound-sa-token\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.942912 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9e0d9776-fb39-4996-9d7d-48a045cd2da9-registry-tls\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.942935 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqwrj\" (UniqueName: \"kubernetes.io/projected/182965cc-4f7e-47e6-a33d-55aa53279d45-kube-api-access-xqwrj\") pod \"machine-config-operator-74547568cd-xwszv\" (UID: \"182965cc-4f7e-47e6-a33d-55aa53279d45\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.942969 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e0d9776-fb39-4996-9d7d-48a045cd2da9-trusted-ca\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.943015 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/182965cc-4f7e-47e6-a33d-55aa53279d45-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xwszv\" (UID: \"182965cc-4f7e-47e6-a33d-55aa53279d45\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.943075 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9e0d9776-fb39-4996-9d7d-48a045cd2da9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.943608 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/182965cc-4f7e-47e6-a33d-55aa53279d45-images\") pod \"machine-config-operator-74547568cd-xwszv\" (UID: \"182965cc-4f7e-47e6-a33d-55aa53279d45\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.943626 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/182965cc-4f7e-47e6-a33d-55aa53279d45-proxy-tls\") pod \"machine-config-operator-74547568cd-xwszv\" (UID: \"182965cc-4f7e-47e6-a33d-55aa53279d45\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv" Feb 22 00:08:54 crc kubenswrapper[4857]: E0222 00:08:54.945875 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:55.445862874 +0000 UTC m=+143.084592127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.958555 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.968621 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4drbd" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.973721 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-99cq6" Feb 22 00:08:54 crc kubenswrapper[4857]: I0222 00:08:54.997549 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.008479 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.012115 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.031340 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tctk6" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.036636 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-h8pq2" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.044982 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:55 crc kubenswrapper[4857]: E0222 00:08:55.045132 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:55.545110324 +0000 UTC m=+143.183839577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.046518 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.046576 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mjr6\" (UniqueName: \"kubernetes.io/projected/9e0d9776-fb39-4996-9d7d-48a045cd2da9-kube-api-access-6mjr6\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.046596 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.046632 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/34787fe4-1fcc-4997-a612-f787f990c3d8-apiservice-cert\") pod \"packageserver-d55dfcdfc-tm4ld\" (UID: \"34787fe4-1fcc-4997-a612-f787f990c3d8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.046665 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqt9l\" (UniqueName: \"kubernetes.io/projected/e43cf3d5-a9c1-4804-a7cd-d0276c875daf-kube-api-access-qqt9l\") pod \"dns-default-c4xt9\" (UID: \"e43cf3d5-a9c1-4804-a7cd-d0276c875daf\") " pod="openshift-dns/dns-default-c4xt9" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.046718 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/34787fe4-1fcc-4997-a612-f787f990c3d8-tmpfs\") pod \"packageserver-d55dfcdfc-tm4ld\" (UID: \"34787fe4-1fcc-4997-a612-f787f990c3d8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.046733 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d89916ae-f616-40ee-af76-fe17a677a488-cert\") pod \"ingress-canary-p9ztw\" (UID: \"d89916ae-f616-40ee-af76-fe17a677a488\") " pod="openshift-ingress-canary/ingress-canary-p9ztw" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.046749 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ad13880d-4f52-4330-a02b-1f3006a9d979-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hm6qc\" (UID: \"ad13880d-4f52-4330-a02b-1f3006a9d979\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.046776 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/46391957-feff-4548-8c20-c713712585af-proxy-tls\") pod \"machine-config-controller-84d6567774-6xx4f\" (UID: \"46391957-feff-4548-8c20-c713712585af\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6xx4f" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.046792 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9e0d9776-fb39-4996-9d7d-48a045cd2da9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.046807 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.046865 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.046883 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.046919 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/46391957-feff-4548-8c20-c713712585af-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6xx4f\" (UID: \"46391957-feff-4548-8c20-c713712585af\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6xx4f" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.046936 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7b02ac9a-c64c-41aa-b2ef-dbff783bf797-mountpoint-dir\") pod \"csi-hostpathplugin-9x24j\" (UID: \"7b02ac9a-c64c-41aa-b2ef-dbff783bf797\") " pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.046952 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.046972 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047051 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vmk2g\" (UID: \"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8\") " pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047098 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5nh9\" (UniqueName: \"kubernetes.io/projected/46391957-feff-4548-8c20-c713712585af-kube-api-access-p5nh9\") pod \"machine-config-controller-84d6567774-6xx4f\" (UID: \"46391957-feff-4548-8c20-c713712585af\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6xx4f" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047119 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpdks\" (UniqueName: \"kubernetes.io/projected/816930b0-fec7-4eb4-bca1-32ca5775f5be-kube-api-access-fpdks\") pod \"machine-config-server-f6z92\" (UID: \"816930b0-fec7-4eb4-bca1-32ca5775f5be\") " pod="openshift-machine-config-operator/machine-config-server-f6z92" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047137 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9e0d9776-fb39-4996-9d7d-48a045cd2da9-registry-certificates\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047222 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7b02ac9a-c64c-41aa-b2ef-dbff783bf797-plugins-dir\") pod \"csi-hostpathplugin-9x24j\" (UID: \"7b02ac9a-c64c-41aa-b2ef-dbff783bf797\") " pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047241 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kg5t\" (UniqueName: \"kubernetes.io/projected/e0b1777b-2a7f-4b10-8051-bb35cb18db72-kube-api-access-5kg5t\") pod \"service-ca-operator-777779d784-d4kmf\" (UID: \"e0b1777b-2a7f-4b10-8051-bb35cb18db72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4kmf" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047265 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9e0d9776-fb39-4996-9d7d-48a045cd2da9-bound-sa-token\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047280 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rczr6\" (UniqueName: \"kubernetes.io/projected/d89916ae-f616-40ee-af76-fe17a677a488-kube-api-access-rczr6\") pod \"ingress-canary-p9ztw\" (UID: \"d89916ae-f616-40ee-af76-fe17a677a488\") " pod="openshift-ingress-canary/ingress-canary-p9ztw" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047296 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047314 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7b02ac9a-c64c-41aa-b2ef-dbff783bf797-registration-dir\") pod \"csi-hostpathplugin-9x24j\" (UID: \"7b02ac9a-c64c-41aa-b2ef-dbff783bf797\") " pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047336 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk5ml\" (UniqueName: \"kubernetes.io/projected/34787fe4-1fcc-4997-a612-f787f990c3d8-kube-api-access-kk5ml\") pod \"packageserver-d55dfcdfc-tm4ld\" (UID: \"34787fe4-1fcc-4997-a612-f787f990c3d8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047358 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7b02ac9a-c64c-41aa-b2ef-dbff783bf797-csi-data-dir\") pod \"csi-hostpathplugin-9x24j\" (UID: \"7b02ac9a-c64c-41aa-b2ef-dbff783bf797\") " pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047397 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7b02ac9a-c64c-41aa-b2ef-dbff783bf797-socket-dir\") pod \"csi-hostpathplugin-9x24j\" (UID: \"7b02ac9a-c64c-41aa-b2ef-dbff783bf797\") " pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047416 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/34787fe4-1fcc-4997-a612-f787f990c3d8-webhook-cert\") pod \"packageserver-d55dfcdfc-tm4ld\" (UID: \"34787fe4-1fcc-4997-a612-f787f990c3d8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047433 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vmk2g\" (UID: \"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8\") " pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047452 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ad13880d-4f52-4330-a02b-1f3006a9d979-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hm6qc\" (UID: \"ad13880d-4f52-4330-a02b-1f3006a9d979\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047467 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e43cf3d5-a9c1-4804-a7cd-d0276c875daf-config-volume\") pod \"dns-default-c4xt9\" (UID: \"e43cf3d5-a9c1-4804-a7cd-d0276c875daf\") " pod="openshift-dns/dns-default-c4xt9" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047500 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-m6mzr\" (UID: \"0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-m6mzr" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047516 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ad13880d-4f52-4330-a02b-1f3006a9d979-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hm6qc\" (UID: \"ad13880d-4f52-4330-a02b-1f3006a9d979\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047578 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjjrg\" (UniqueName: \"kubernetes.io/projected/7b02ac9a-c64c-41aa-b2ef-dbff783bf797-kube-api-access-fjjrg\") pod \"csi-hostpathplugin-9x24j\" (UID: \"7b02ac9a-c64c-41aa-b2ef-dbff783bf797\") " pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047613 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9e0d9776-fb39-4996-9d7d-48a045cd2da9-registry-tls\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047638 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqwrj\" (UniqueName: \"kubernetes.io/projected/182965cc-4f7e-47e6-a33d-55aa53279d45-kube-api-access-xqwrj\") pod \"machine-config-operator-74547568cd-xwszv\" (UID: \"182965cc-4f7e-47e6-a33d-55aa53279d45\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047654 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/816930b0-fec7-4eb4-bca1-32ca5775f5be-certs\") pod \"machine-config-server-f6z92\" (UID: \"816930b0-fec7-4eb4-bca1-32ca5775f5be\") " pod="openshift-machine-config-operator/machine-config-server-f6z92" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047687 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047705 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sj6pj\" (UniqueName: \"kubernetes.io/projected/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-kube-api-access-sj6pj\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047778 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kqck\" (UniqueName: \"kubernetes.io/projected/ad13880d-4f52-4330-a02b-1f3006a9d979-kube-api-access-8kqck\") pod \"cluster-image-registry-operator-dc59b4c8b-hm6qc\" (UID: \"ad13880d-4f52-4330-a02b-1f3006a9d979\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047873 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047903 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/816930b0-fec7-4eb4-bca1-32ca5775f5be-node-bootstrap-token\") pod \"machine-config-server-f6z92\" (UID: \"816930b0-fec7-4eb4-bca1-32ca5775f5be\") " pod="openshift-machine-config-operator/machine-config-server-f6z92" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047941 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-audit-policies\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.047994 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e0d9776-fb39-4996-9d7d-48a045cd2da9-trusted-ca\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.048146 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-m6mzr\" (UID: \"0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-m6mzr" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.048172 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-audit-dir\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.048196 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/182965cc-4f7e-47e6-a33d-55aa53279d45-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xwszv\" (UID: \"182965cc-4f7e-47e6-a33d-55aa53279d45\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.048229 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.048284 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.048320 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtzzn\" (UniqueName: \"kubernetes.io/projected/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8-kube-api-access-qtzzn\") pod \"marketplace-operator-79b997595-vmk2g\" (UID: \"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8\") " pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.048337 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n4d9\" (UniqueName: \"kubernetes.io/projected/0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6-kube-api-access-5n4d9\") pod \"kube-storage-version-migrator-operator-b67b599dd-m6mzr\" (UID: \"0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-m6mzr" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.048395 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0b1777b-2a7f-4b10-8051-bb35cb18db72-serving-cert\") pod \"service-ca-operator-777779d784-d4kmf\" (UID: \"e0b1777b-2a7f-4b10-8051-bb35cb18db72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4kmf" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.048431 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9e0d9776-fb39-4996-9d7d-48a045cd2da9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.048458 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/182965cc-4f7e-47e6-a33d-55aa53279d45-images\") pod \"machine-config-operator-74547568cd-xwszv\" (UID: \"182965cc-4f7e-47e6-a33d-55aa53279d45\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.048473 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/182965cc-4f7e-47e6-a33d-55aa53279d45-proxy-tls\") pod \"machine-config-operator-74547568cd-xwszv\" (UID: \"182965cc-4f7e-47e6-a33d-55aa53279d45\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.048490 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e43cf3d5-a9c1-4804-a7cd-d0276c875daf-metrics-tls\") pod \"dns-default-c4xt9\" (UID: \"e43cf3d5-a9c1-4804-a7cd-d0276c875daf\") " pod="openshift-dns/dns-default-c4xt9" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.048535 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0b1777b-2a7f-4b10-8051-bb35cb18db72-config\") pod \"service-ca-operator-777779d784-d4kmf\" (UID: \"e0b1777b-2a7f-4b10-8051-bb35cb18db72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4kmf" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.050195 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/182965cc-4f7e-47e6-a33d-55aa53279d45-images\") pod \"machine-config-operator-74547568cd-xwszv\" (UID: \"182965cc-4f7e-47e6-a33d-55aa53279d45\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv" Feb 22 00:08:55 crc kubenswrapper[4857]: E0222 00:08:55.050267 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:55.55025158 +0000 UTC m=+143.188980833 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.052251 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/182965cc-4f7e-47e6-a33d-55aa53279d45-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xwszv\" (UID: \"182965cc-4f7e-47e6-a33d-55aa53279d45\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.052432 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/46391957-feff-4548-8c20-c713712585af-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-6xx4f\" (UID: \"46391957-feff-4548-8c20-c713712585af\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6xx4f" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.053230 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e0d9776-fb39-4996-9d7d-48a045cd2da9-trusted-ca\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.053771 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9e0d9776-fb39-4996-9d7d-48a045cd2da9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.054001 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9e0d9776-fb39-4996-9d7d-48a045cd2da9-registry-certificates\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.064341 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9e0d9776-fb39-4996-9d7d-48a045cd2da9-registry-tls\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.064407 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-ctjrv"] Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.065931 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/182965cc-4f7e-47e6-a33d-55aa53279d45-proxy-tls\") pod \"machine-config-operator-74547568cd-xwszv\" (UID: \"182965cc-4f7e-47e6-a33d-55aa53279d45\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.067666 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/46391957-feff-4548-8c20-c713712585af-proxy-tls\") pod \"machine-config-controller-84d6567774-6xx4f\" (UID: \"46391957-feff-4548-8c20-c713712585af\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6xx4f" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.070249 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9e0d9776-fb39-4996-9d7d-48a045cd2da9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.100017 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9e0d9776-fb39-4996-9d7d-48a045cd2da9-bound-sa-token\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.118271 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mjr6\" (UniqueName: \"kubernetes.io/projected/9e0d9776-fb39-4996-9d7d-48a045cd2da9-kube-api-access-6mjr6\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.147928 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6d9wr"] Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.150445 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4nvsx"] Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151258 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151444 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151475 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtzzn\" (UniqueName: \"kubernetes.io/projected/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8-kube-api-access-qtzzn\") pod \"marketplace-operator-79b997595-vmk2g\" (UID: \"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8\") " pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151490 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n4d9\" (UniqueName: \"kubernetes.io/projected/0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6-kube-api-access-5n4d9\") pod \"kube-storage-version-migrator-operator-b67b599dd-m6mzr\" (UID: \"0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-m6mzr" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151508 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0b1777b-2a7f-4b10-8051-bb35cb18db72-serving-cert\") pod \"service-ca-operator-777779d784-d4kmf\" (UID: \"e0b1777b-2a7f-4b10-8051-bb35cb18db72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4kmf" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151530 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e43cf3d5-a9c1-4804-a7cd-d0276c875daf-metrics-tls\") pod \"dns-default-c4xt9\" (UID: \"e43cf3d5-a9c1-4804-a7cd-d0276c875daf\") " pod="openshift-dns/dns-default-c4xt9" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151554 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0b1777b-2a7f-4b10-8051-bb35cb18db72-config\") pod \"service-ca-operator-777779d784-d4kmf\" (UID: \"e0b1777b-2a7f-4b10-8051-bb35cb18db72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4kmf" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151571 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151595 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151614 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/34787fe4-1fcc-4997-a612-f787f990c3d8-apiservice-cert\") pod \"packageserver-d55dfcdfc-tm4ld\" (UID: \"34787fe4-1fcc-4997-a612-f787f990c3d8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151630 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqt9l\" (UniqueName: \"kubernetes.io/projected/e43cf3d5-a9c1-4804-a7cd-d0276c875daf-kube-api-access-qqt9l\") pod \"dns-default-c4xt9\" (UID: \"e43cf3d5-a9c1-4804-a7cd-d0276c875daf\") " pod="openshift-dns/dns-default-c4xt9" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151652 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/34787fe4-1fcc-4997-a612-f787f990c3d8-tmpfs\") pod \"packageserver-d55dfcdfc-tm4ld\" (UID: \"34787fe4-1fcc-4997-a612-f787f990c3d8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151665 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d89916ae-f616-40ee-af76-fe17a677a488-cert\") pod \"ingress-canary-p9ztw\" (UID: \"d89916ae-f616-40ee-af76-fe17a677a488\") " pod="openshift-ingress-canary/ingress-canary-p9ztw" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151687 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ad13880d-4f52-4330-a02b-1f3006a9d979-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hm6qc\" (UID: \"ad13880d-4f52-4330-a02b-1f3006a9d979\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151704 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151722 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151738 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151755 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7b02ac9a-c64c-41aa-b2ef-dbff783bf797-mountpoint-dir\") pod \"csi-hostpathplugin-9x24j\" (UID: \"7b02ac9a-c64c-41aa-b2ef-dbff783bf797\") " pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151771 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151802 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vmk2g\" (UID: \"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8\") " pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151825 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpdks\" (UniqueName: \"kubernetes.io/projected/816930b0-fec7-4eb4-bca1-32ca5775f5be-kube-api-access-fpdks\") pod \"machine-config-server-f6z92\" (UID: \"816930b0-fec7-4eb4-bca1-32ca5775f5be\") " pod="openshift-machine-config-operator/machine-config-server-f6z92" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151850 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7b02ac9a-c64c-41aa-b2ef-dbff783bf797-plugins-dir\") pod \"csi-hostpathplugin-9x24j\" (UID: \"7b02ac9a-c64c-41aa-b2ef-dbff783bf797\") " pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151864 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kg5t\" (UniqueName: \"kubernetes.io/projected/e0b1777b-2a7f-4b10-8051-bb35cb18db72-kube-api-access-5kg5t\") pod \"service-ca-operator-777779d784-d4kmf\" (UID: \"e0b1777b-2a7f-4b10-8051-bb35cb18db72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4kmf" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151879 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczr6\" (UniqueName: \"kubernetes.io/projected/d89916ae-f616-40ee-af76-fe17a677a488-kube-api-access-rczr6\") pod \"ingress-canary-p9ztw\" (UID: \"d89916ae-f616-40ee-af76-fe17a677a488\") " pod="openshift-ingress-canary/ingress-canary-p9ztw" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151894 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151909 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7b02ac9a-c64c-41aa-b2ef-dbff783bf797-registration-dir\") pod \"csi-hostpathplugin-9x24j\" (UID: \"7b02ac9a-c64c-41aa-b2ef-dbff783bf797\") " pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151925 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk5ml\" (UniqueName: \"kubernetes.io/projected/34787fe4-1fcc-4997-a612-f787f990c3d8-kube-api-access-kk5ml\") pod \"packageserver-d55dfcdfc-tm4ld\" (UID: \"34787fe4-1fcc-4997-a612-f787f990c3d8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151939 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7b02ac9a-c64c-41aa-b2ef-dbff783bf797-csi-data-dir\") pod \"csi-hostpathplugin-9x24j\" (UID: \"7b02ac9a-c64c-41aa-b2ef-dbff783bf797\") " pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151956 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7b02ac9a-c64c-41aa-b2ef-dbff783bf797-socket-dir\") pod \"csi-hostpathplugin-9x24j\" (UID: \"7b02ac9a-c64c-41aa-b2ef-dbff783bf797\") " pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151972 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/34787fe4-1fcc-4997-a612-f787f990c3d8-webhook-cert\") pod \"packageserver-d55dfcdfc-tm4ld\" (UID: \"34787fe4-1fcc-4997-a612-f787f990c3d8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.151986 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vmk2g\" (UID: \"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8\") " pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.152000 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ad13880d-4f52-4330-a02b-1f3006a9d979-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hm6qc\" (UID: \"ad13880d-4f52-4330-a02b-1f3006a9d979\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.152014 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e43cf3d5-a9c1-4804-a7cd-d0276c875daf-config-volume\") pod \"dns-default-c4xt9\" (UID: \"e43cf3d5-a9c1-4804-a7cd-d0276c875daf\") " pod="openshift-dns/dns-default-c4xt9" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.154144 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: E0222 00:08:55.154226 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:55.654214284 +0000 UTC m=+143.292943527 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.154828 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.155667 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-m6mzr\" (UID: \"0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-m6mzr" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.155897 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ad13880d-4f52-4330-a02b-1f3006a9d979-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hm6qc\" (UID: \"ad13880d-4f52-4330-a02b-1f3006a9d979\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.155931 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjjrg\" (UniqueName: \"kubernetes.io/projected/7b02ac9a-c64c-41aa-b2ef-dbff783bf797-kube-api-access-fjjrg\") pod \"csi-hostpathplugin-9x24j\" (UID: \"7b02ac9a-c64c-41aa-b2ef-dbff783bf797\") " pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.156015 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/7b02ac9a-c64c-41aa-b2ef-dbff783bf797-plugins-dir\") pod \"csi-hostpathplugin-9x24j\" (UID: \"7b02ac9a-c64c-41aa-b2ef-dbff783bf797\") " pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.156880 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/816930b0-fec7-4eb4-bca1-32ca5775f5be-certs\") pod \"machine-config-server-f6z92\" (UID: \"816930b0-fec7-4eb4-bca1-32ca5775f5be\") " pod="openshift-machine-config-operator/machine-config-server-f6z92" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.156909 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.156909 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0b1777b-2a7f-4b10-8051-bb35cb18db72-config\") pod \"service-ca-operator-777779d784-d4kmf\" (UID: \"e0b1777b-2a7f-4b10-8051-bb35cb18db72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4kmf" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.156930 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sj6pj\" (UniqueName: \"kubernetes.io/projected/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-kube-api-access-sj6pj\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.156975 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kqck\" (UniqueName: \"kubernetes.io/projected/ad13880d-4f52-4330-a02b-1f3006a9d979-kube-api-access-8kqck\") pod \"cluster-image-registry-operator-dc59b4c8b-hm6qc\" (UID: \"ad13880d-4f52-4330-a02b-1f3006a9d979\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.157024 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.157068 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/816930b0-fec7-4eb4-bca1-32ca5775f5be-node-bootstrap-token\") pod \"machine-config-server-f6z92\" (UID: \"816930b0-fec7-4eb4-bca1-32ca5775f5be\") " pod="openshift-machine-config-operator/machine-config-server-f6z92" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.157131 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-audit-policies\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.157189 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-m6mzr\" (UID: \"0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-m6mzr" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.157214 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-audit-dir\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.157248 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.157395 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ad13880d-4f52-4330-a02b-1f3006a9d979-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hm6qc\" (UID: \"ad13880d-4f52-4330-a02b-1f3006a9d979\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.157467 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/7b02ac9a-c64c-41aa-b2ef-dbff783bf797-mountpoint-dir\") pod \"csi-hostpathplugin-9x24j\" (UID: \"7b02ac9a-c64c-41aa-b2ef-dbff783bf797\") " pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.158487 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/7b02ac9a-c64c-41aa-b2ef-dbff783bf797-registration-dir\") pod \"csi-hostpathplugin-9x24j\" (UID: \"7b02ac9a-c64c-41aa-b2ef-dbff783bf797\") " pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.158649 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/7b02ac9a-c64c-41aa-b2ef-dbff783bf797-csi-data-dir\") pod \"csi-hostpathplugin-9x24j\" (UID: \"7b02ac9a-c64c-41aa-b2ef-dbff783bf797\") " pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.158716 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/7b02ac9a-c64c-41aa-b2ef-dbff783bf797-socket-dir\") pod \"csi-hostpathplugin-9x24j\" (UID: \"7b02ac9a-c64c-41aa-b2ef-dbff783bf797\") " pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.156972 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/34787fe4-1fcc-4997-a612-f787f990c3d8-tmpfs\") pod \"packageserver-d55dfcdfc-tm4ld\" (UID: \"34787fe4-1fcc-4997-a612-f787f990c3d8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.159721 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vmk2g\" (UID: \"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8\") " pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.159882 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-audit-dir\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.160165 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e43cf3d5-a9c1-4804-a7cd-d0276c875daf-config-volume\") pod \"dns-default-c4xt9\" (UID: \"e43cf3d5-a9c1-4804-a7cd-d0276c875daf\") " pod="openshift-dns/dns-default-c4xt9" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.160290 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-audit-policies\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.160311 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-m6mzr\" (UID: \"0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-m6mzr" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.160934 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.162294 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.164591 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/816930b0-fec7-4eb4-bca1-32ca5775f5be-certs\") pod \"machine-config-server-f6z92\" (UID: \"816930b0-fec7-4eb4-bca1-32ca5775f5be\") " pod="openshift-machine-config-operator/machine-config-server-f6z92" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.164646 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.164852 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e43cf3d5-a9c1-4804-a7cd-d0276c875daf-metrics-tls\") pod \"dns-default-c4xt9\" (UID: \"e43cf3d5-a9c1-4804-a7cd-d0276c875daf\") " pod="openshift-dns/dns-default-c4xt9" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.165025 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqwrj\" (UniqueName: \"kubernetes.io/projected/182965cc-4f7e-47e6-a33d-55aa53279d45-kube-api-access-xqwrj\") pod \"machine-config-operator-74547568cd-xwszv\" (UID: \"182965cc-4f7e-47e6-a33d-55aa53279d45\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.165453 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d89916ae-f616-40ee-af76-fe17a677a488-cert\") pod \"ingress-canary-p9ztw\" (UID: \"d89916ae-f616-40ee-af76-fe17a677a488\") " pod="openshift-ingress-canary/ingress-canary-p9ztw" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.165884 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/816930b0-fec7-4eb4-bca1-32ca5775f5be-node-bootstrap-token\") pod \"machine-config-server-f6z92\" (UID: \"816930b0-fec7-4eb4-bca1-32ca5775f5be\") " pod="openshift-machine-config-operator/machine-config-server-f6z92" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.189471 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.189800 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.189827 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.189980 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/34787fe4-1fcc-4997-a612-f787f990c3d8-apiservice-cert\") pod \"packageserver-d55dfcdfc-tm4ld\" (UID: \"34787fe4-1fcc-4997-a612-f787f990c3d8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.190221 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0b1777b-2a7f-4b10-8051-bb35cb18db72-serving-cert\") pod \"service-ca-operator-777779d784-d4kmf\" (UID: \"e0b1777b-2a7f-4b10-8051-bb35cb18db72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4kmf" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.190431 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ad13880d-4f52-4330-a02b-1f3006a9d979-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hm6qc\" (UID: \"ad13880d-4f52-4330-a02b-1f3006a9d979\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.190464 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5nh9\" (UniqueName: \"kubernetes.io/projected/46391957-feff-4548-8c20-c713712585af-kube-api-access-p5nh9\") pod \"machine-config-controller-84d6567774-6xx4f\" (UID: \"46391957-feff-4548-8c20-c713712585af\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6xx4f" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.193952 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wq8b8"] Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.194892 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29528640-g7sjv"] Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.197437 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.197648 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-m6mzr\" (UID: \"0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-m6mzr" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.199687 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.206686 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/34787fe4-1fcc-4997-a612-f787f990c3d8-webhook-cert\") pod \"packageserver-d55dfcdfc-tm4ld\" (UID: \"34787fe4-1fcc-4997-a612-f787f990c3d8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.207146 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vmk2g\" (UID: \"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8\") " pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.207467 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.209270 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6xx4f" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.210027 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpdks\" (UniqueName: \"kubernetes.io/projected/816930b0-fec7-4eb4-bca1-32ca5775f5be-kube-api-access-fpdks\") pod \"machine-config-server-f6z92\" (UID: \"816930b0-fec7-4eb4-bca1-32ca5775f5be\") " pod="openshift-machine-config-operator/machine-config-server-f6z92" Feb 22 00:08:55 crc kubenswrapper[4857]: W0222 00:08:55.211084 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a31ed34_05d3_4e0f_bb4b_03d086e01f84.slice/crio-9712a2416c52f029beef5deb2f3c28bb760cdc4480f7bccb6e002cf62bf9fa4b WatchSource:0}: Error finding container 9712a2416c52f029beef5deb2f3c28bb760cdc4480f7bccb6e002cf62bf9fa4b: Status 404 returned error can't find the container with id 9712a2416c52f029beef5deb2f3c28bb760cdc4480f7bccb6e002cf62bf9fa4b Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.214633 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kg5t\" (UniqueName: \"kubernetes.io/projected/e0b1777b-2a7f-4b10-8051-bb35cb18db72-kube-api-access-5kg5t\") pod \"service-ca-operator-777779d784-d4kmf\" (UID: \"e0b1777b-2a7f-4b10-8051-bb35cb18db72\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4kmf" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.228701 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-cwlf7"] Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.234523 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sj6pj\" (UniqueName: \"kubernetes.io/projected/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-kube-api-access-sj6pj\") pod \"oauth-openshift-558db77b4-lm22m\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.256637 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.257751 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: E0222 00:08:55.258056 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:55.758033574 +0000 UTC m=+143.396762827 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.278979 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n4d9\" (UniqueName: \"kubernetes.io/projected/0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6-kube-api-access-5n4d9\") pod \"kube-storage-version-migrator-operator-b67b599dd-m6mzr\" (UID: \"0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-m6mzr" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.297123 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtzzn\" (UniqueName: \"kubernetes.io/projected/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8-kube-api-access-qtzzn\") pod \"marketplace-operator-79b997595-vmk2g\" (UID: \"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8\") " pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.299835 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6vrqd"] Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.299992 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.304888 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-m6mzr" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.306476 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk9zf"] Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.308019 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-8sts2"] Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.315862 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjjrg\" (UniqueName: \"kubernetes.io/projected/7b02ac9a-c64c-41aa-b2ef-dbff783bf797-kube-api-access-fjjrg\") pod \"csi-hostpathplugin-9x24j\" (UID: \"7b02ac9a-c64c-41aa-b2ef-dbff783bf797\") " pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.321849 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4kmf" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.336018 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk5ml\" (UniqueName: \"kubernetes.io/projected/34787fe4-1fcc-4997-a612-f787f990c3d8-kube-api-access-kk5ml\") pod \"packageserver-d55dfcdfc-tm4ld\" (UID: \"34787fe4-1fcc-4997-a612-f787f990c3d8\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.347139 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqt9l\" (UniqueName: \"kubernetes.io/projected/e43cf3d5-a9c1-4804-a7cd-d0276c875daf-kube-api-access-qqt9l\") pod \"dns-default-c4xt9\" (UID: \"e43cf3d5-a9c1-4804-a7cd-d0276c875daf\") " pod="openshift-dns/dns-default-c4xt9" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.354692 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-f6z92" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.358535 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:55 crc kubenswrapper[4857]: E0222 00:08:55.358687 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:55.858664963 +0000 UTC m=+143.497394226 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.359017 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: E0222 00:08:55.359676 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:55.859667471 +0000 UTC m=+143.498396724 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.379521 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-h4pmt"] Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.385125 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-9x24j" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.389706 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczr6\" (UniqueName: \"kubernetes.io/projected/d89916ae-f616-40ee-af76-fe17a677a488-kube-api-access-rczr6\") pod \"ingress-canary-p9ztw\" (UID: \"d89916ae-f616-40ee-af76-fe17a677a488\") " pod="openshift-ingress-canary/ingress-canary-p9ztw" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.395337 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv"] Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.399022 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-mx94d"] Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.410856 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kqck\" (UniqueName: \"kubernetes.io/projected/ad13880d-4f52-4330-a02b-1f3006a9d979-kube-api-access-8kqck\") pod \"cluster-image-registry-operator-dc59b4c8b-hm6qc\" (UID: \"ad13880d-4f52-4330-a02b-1f3006a9d979\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.411525 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ad13880d-4f52-4330-a02b-1f3006a9d979-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hm6qc\" (UID: \"ad13880d-4f52-4330-a02b-1f3006a9d979\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.461463 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:55 crc kubenswrapper[4857]: E0222 00:08:55.461695 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:55.96166481 +0000 UTC m=+143.600394063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.461875 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: E0222 00:08:55.462235 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:55.962220336 +0000 UTC m=+143.600949579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:55 crc kubenswrapper[4857]: W0222 00:08:55.473001 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod816930b0_fec7_4eb4_bca1_32ca5775f5be.slice/crio-2b510fe3ba6651c8df3257bef37189d331567812564cb71da7e76ebe03ef5f55 WatchSource:0}: Error finding container 2b510fe3ba6651c8df3257bef37189d331567812564cb71da7e76ebe03ef5f55: Status 404 returned error can't find the container with id 2b510fe3ba6651c8df3257bef37189d331567812564cb71da7e76ebe03ef5f55 Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.549542 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.550079 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.550377 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.562822 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:55 crc kubenswrapper[4857]: E0222 00:08:55.563172 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:56.063128803 +0000 UTC m=+143.701858056 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.563216 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: E0222 00:08:55.563565 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:56.063552125 +0000 UTC m=+143.702281378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.584599 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2p58w"] Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.644460 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-c4xt9" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.666344 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-p9ztw" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.666857 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:55 crc kubenswrapper[4857]: E0222 00:08:55.667836 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:56.167804606 +0000 UTC m=+143.806533859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.768614 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: E0222 00:08:55.769026 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:56.269011342 +0000 UTC m=+143.907740595 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.837880 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6d9wr" event={"ID":"0fa7f27c-4fa5-45db-b376-99c5c6468ec2","Type":"ContainerStarted","Data":"9d6293fe89d7d977a421160b29f922e3296a39fbdfff64f7fbcf40a5ae1ab54a"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.840602 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f" event={"ID":"4edfa330-6a4a-4f86-b606-db0adb8301c8","Type":"ContainerStarted","Data":"9724188b6621a8c43adff63682b6166c7ae441d97984480970b83567af7938dd"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.857068 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6vrqd" event={"ID":"5fb885c7-2a02-4adb-af59-35d8d9a1a3d9","Type":"ContainerStarted","Data":"1c1c065a62fe40092705ba856de8370acf98777c16f9009a957319961367c61f"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.858444 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lfd8v" event={"ID":"1d0786e3-c21c-46f5-8ca9-7cad8ac62194","Type":"ContainerStarted","Data":"a0a4b719f666f93f88acfc92400c501a8ccacf2c14fad01f19e8ef0d91fc999c"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.861782 4857 generic.go:334] "Generic (PLEG): container finished" podID="56f6f504-2f13-4ae2-b5cd-d67ec7c140a7" containerID="b36be81daf0238605453a7eb9a79cbec40c5f8dead588786d55f68ba40fb2ab4" exitCode=0 Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.862048 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" event={"ID":"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7","Type":"ContainerDied","Data":"b36be81daf0238605453a7eb9a79cbec40c5f8dead588786d55f68ba40fb2ab4"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.863515 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29528640-g7sjv" event={"ID":"f07c5f6c-743d-4032-b70b-ecc0f514295d","Type":"ContainerStarted","Data":"4bcb8686a27b403e0c8cbe7bc18c2a7f298997eae2c838b76ca4e10000d0154e"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.864614 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv" event={"ID":"706cd825-7128-4833-8260-b3b73cccaac7","Type":"ContainerStarted","Data":"188cf3b0fbf99e96cfde8aa96271efe8ec16466bc6548f755a982ac0efd2c085"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.865635 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" event={"ID":"5a31ed34-05d3-4e0f-bb4b-03d086e01f84","Type":"ContainerStarted","Data":"9712a2416c52f029beef5deb2f3c28bb760cdc4480f7bccb6e002cf62bf9fa4b"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.868572 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-f6z92" event={"ID":"816930b0-fec7-4eb4-bca1-32ca5775f5be","Type":"ContainerStarted","Data":"2b510fe3ba6651c8df3257bef37189d331567812564cb71da7e76ebe03ef5f55"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.869359 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:55 crc kubenswrapper[4857]: E0222 00:08:55.869495 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:56.369478476 +0000 UTC m=+144.008207729 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.869623 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:55 crc kubenswrapper[4857]: E0222 00:08:55.870419 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:56.370402613 +0000 UTC m=+144.009131866 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.884978 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4nvsx" event={"ID":"e00acf3c-29bc-447d-bb80-7a1a8bf7cd84","Type":"ContainerStarted","Data":"21392f2d8d905cead677086a89b7966e4e87cf0cd978e3e3595037bbd3c1fff8"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.890299 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g" event={"ID":"c719ea96-d4bf-4f14-b8bd-c32bb6a20c46","Type":"ContainerStarted","Data":"080cbf421b75fff31f95b3ac692507360d2e90ba2b308934b5e8ff7b88cd47d2"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.906916 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2p58w" event={"ID":"61fba77d-c313-4c67-a0d8-7fbf2e172aa4","Type":"ContainerStarted","Data":"5760b83791c44be7e000c568db4bdda70593a3bd1a3a1f73912657b4b6fedea3"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.911533 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-cwlf7" event={"ID":"f9bce83c-ae40-482f-88d2-7de36a1a7ea5","Type":"ContainerStarted","Data":"1ba661b6e9efc8c2112732faf19c36c57faf64f7dc31e62df3af94f1354a73a0"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.914131 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nj6dw" event={"ID":"c2b40552-4007-4981-ba86-2d8864fc20cf","Type":"ContainerStarted","Data":"39c6c4f69b6887c0883321c36ec6958156efe3601069c592d1eeb665eb92ce0c"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.916810 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wq8b8" event={"ID":"f5acab23-9994-4c18-bc50-4f5ec7c7f525","Type":"ContainerStarted","Data":"c5a527919a3bbfe7f3a0a5042f1db8dff84b13c22d370917dc9f063779d8bf80"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.920867 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-h4pmt" event={"ID":"29721fdf-b040-414a-a6ca-6ba9dc544526","Type":"ContainerStarted","Data":"2a4c576dee6996f3eaa8b9f56fe67b20dc8ea88f87fcc0f89d360206a480cf68"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.944246 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" event={"ID":"fef422b7-19df-431d-b574-73457050f022","Type":"ContainerStarted","Data":"b99fab72ef39a65c6ea69cdcf8740df7fe6fa7340e2b33bbfb638f351cb6aee9"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.967451 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" event={"ID":"e8326fb9-1a3f-402a-acf2-62f7d6e82411","Type":"ContainerStarted","Data":"b967b3799020d96b36d89771640a99805c2f6655ec137db205835b0199ff5a12"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.968190 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.971398 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:55 crc kubenswrapper[4857]: E0222 00:08:55.972351 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:56.472335429 +0000 UTC m=+144.111064682 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.986582 4857 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-c669b container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.986636 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" podUID="e8326fb9-1a3f-402a-acf2-62f7d6e82411" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.11:8443/healthz\": dial tcp 10.217.0.11:8443: connect: connection refused" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.990328 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" event={"ID":"60d12108-1ec8-4074-a60c-2b143a83f59f","Type":"ContainerStarted","Data":"7fed6a816beba39e0dfd87ea70ce1f57f5b13bb463fd560b0dc98a55e8cde5b9"} Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.990825 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.995077 4857 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-p979v container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Feb 22 00:08:55 crc kubenswrapper[4857]: I0222 00:08:55.995115 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" podUID="60d12108-1ec8-4074-a60c-2b143a83f59f" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.001707 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-mx94d" event={"ID":"860d5a18-25c6-4484-bc45-2e741cd6b708","Type":"ContainerStarted","Data":"9cca0f6ccc923442a2672cb0d612333282eee340ae99a63b9908aab8e2586a2d"} Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.044255 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8x7sr"] Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.082528 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:56 crc kubenswrapper[4857]: E0222 00:08:56.085199 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:56.585179676 +0000 UTC m=+144.223908929 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.105232 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4drbd"] Feb 22 00:08:56 crc kubenswrapper[4857]: W0222 00:08:56.108251 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6ba7763_9bd5_45dd_a9a9_d47dc7f381ba.slice/crio-c3a03629d1c24a7bbd62d03b3ebceea0b66afadf6309a95ff0053fb61b3703f1 WatchSource:0}: Error finding container c3a03629d1c24a7bbd62d03b3ebceea0b66afadf6309a95ff0053fb61b3703f1: Status 404 returned error can't find the container with id c3a03629d1c24a7bbd62d03b3ebceea0b66afadf6309a95ff0053fb61b3703f1 Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.111914 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c"] Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.120071 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-wc47x"] Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.120716 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g4v7w"] Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.188295 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:56 crc kubenswrapper[4857]: E0222 00:08:56.188609 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:56.688584214 +0000 UTC m=+144.327313467 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.195607 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:56 crc kubenswrapper[4857]: E0222 00:08:56.197391 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:56.697371515 +0000 UTC m=+144.336100768 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.208845 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9"] Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.299626 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:56 crc kubenswrapper[4857]: E0222 00:08:56.299889 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:56.799875098 +0000 UTC m=+144.438604351 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:56 crc kubenswrapper[4857]: W0222 00:08:56.342805 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5c791b9_d379_4311_b961_0927ca48702a.slice/crio-8b014b87ca2a9e635f27dc25c9798639c5547486ed420201dd6e573f46ff41e8 WatchSource:0}: Error finding container 8b014b87ca2a9e635f27dc25c9798639c5547486ed420201dd6e573f46ff41e8: Status 404 returned error can't find the container with id 8b014b87ca2a9e635f27dc25c9798639c5547486ed420201dd6e573f46ff41e8 Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.401532 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:56 crc kubenswrapper[4857]: E0222 00:08:56.401801 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:56.901790563 +0000 UTC m=+144.540519816 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.472770 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-9x24j"] Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.479850 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-tctk6"] Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.504895 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:56 crc kubenswrapper[4857]: E0222 00:08:56.505015 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:57.004998845 +0000 UTC m=+144.643728098 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.505438 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:56 crc kubenswrapper[4857]: E0222 00:08:56.505788 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:57.005779098 +0000 UTC m=+144.644508351 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.519080 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-d4kmf"] Feb 22 00:08:56 crc kubenswrapper[4857]: W0222 00:08:56.535726 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b02ac9a_c64c_41aa_b2ef_dbff783bf797.slice/crio-3828a74d0d1f7de622d84a2997b76075a7be41ed28e2ac675d6e3a3dad835902 WatchSource:0}: Error finding container 3828a74d0d1f7de622d84a2997b76075a7be41ed28e2ac675d6e3a3dad835902: Status 404 returned error can't find the container with id 3828a74d0d1f7de622d84a2997b76075a7be41ed28e2ac675d6e3a3dad835902 Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.546218 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-m6mzr"] Feb 22 00:08:56 crc kubenswrapper[4857]: W0222 00:08:56.549491 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8865dea_cf78_4acd_bcb4_aad234172e99.slice/crio-100a35fe755e4c13c294ea6b29c7e74de3951a004e8f3fd5b1c372bfbc3b5033 WatchSource:0}: Error finding container 100a35fe755e4c13c294ea6b29c7e74de3951a004e8f3fd5b1c372bfbc3b5033: Status 404 returned error can't find the container with id 100a35fe755e4c13c294ea6b29c7e74de3951a004e8f3fd5b1c372bfbc3b5033 Feb 22 00:08:56 crc kubenswrapper[4857]: W0222 00:08:56.549814 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0b1777b_2a7f_4b10_8051_bb35cb18db72.slice/crio-c1e7af8e4033d482ee8273c727e6e1f42e278bddbff0a30d5eced5b825ef477e WatchSource:0}: Error finding container c1e7af8e4033d482ee8273c727e6e1f42e278bddbff0a30d5eced5b825ef477e: Status 404 returned error can't find the container with id c1e7af8e4033d482ee8273c727e6e1f42e278bddbff0a30d5eced5b825ef477e Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.585050 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-99cq6"] Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.614837 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:56 crc kubenswrapper[4857]: E0222 00:08:56.616432 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:57.116413832 +0000 UTC m=+144.755143085 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.630642 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-h8pq2"] Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.637282 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lm22m"] Feb 22 00:08:56 crc kubenswrapper[4857]: W0222 00:08:56.639233 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46391957_feff_4548_8c20_c713712585af.slice/crio-e0aaa06ea1a5dff1a2c4a2fb483e4cf577fc174ccfba80b4501ddf40c1059b6e WatchSource:0}: Error finding container e0aaa06ea1a5dff1a2c4a2fb483e4cf577fc174ccfba80b4501ddf40c1059b6e: Status 404 returned error can't find the container with id e0aaa06ea1a5dff1a2c4a2fb483e4cf577fc174ccfba80b4501ddf40c1059b6e Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.643925 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-6xx4f"] Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.656700 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld"] Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.663158 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv"] Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.687180 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc"] Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.696348 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz"] Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.696471 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-p9ztw"] Feb 22 00:08:56 crc kubenswrapper[4857]: W0222 00:08:56.705271 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e531e33_264a_498e_b279_cf687f1e6327.slice/crio-33cacefea17590c1ae6843b119f7eabcf5c73238554d6c21e4e87b9938f53324 WatchSource:0}: Error finding container 33cacefea17590c1ae6843b119f7eabcf5c73238554d6c21e4e87b9938f53324: Status 404 returned error can't find the container with id 33cacefea17590c1ae6843b119f7eabcf5c73238554d6c21e4e87b9938f53324 Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.717129 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.717312 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-c4xt9"] Feb 22 00:08:56 crc kubenswrapper[4857]: E0222 00:08:56.717471 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:57.217461432 +0000 UTC m=+144.856190685 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.736681 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vmk2g"] Feb 22 00:08:56 crc kubenswrapper[4857]: W0222 00:08:56.736937 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode43cf3d5_a9c1_4804_a7cd_d0276c875daf.slice/crio-44b0e132d2ff4b60d5f7d8f819bacc68a5b6c68ddbd6d1d04641895ec0a31b34 WatchSource:0}: Error finding container 44b0e132d2ff4b60d5f7d8f819bacc68a5b6c68ddbd6d1d04641895ec0a31b34: Status 404 returned error can't find the container with id 44b0e132d2ff4b60d5f7d8f819bacc68a5b6c68ddbd6d1d04641895ec0a31b34 Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.817992 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:56 crc kubenswrapper[4857]: E0222 00:08:56.818439 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:57.318424021 +0000 UTC m=+144.957153274 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:56 crc kubenswrapper[4857]: I0222 00:08:56.919904 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:56 crc kubenswrapper[4857]: E0222 00:08:56.920418 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:57.420401569 +0000 UTC m=+145.059130822 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.020584 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:57 crc kubenswrapper[4857]: E0222 00:08:57.020838 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:57.520824952 +0000 UTC m=+145.159554205 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.041594 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tctk6" event={"ID":"e8865dea-cf78-4acd-bcb4-aad234172e99","Type":"ContainerStarted","Data":"100a35fe755e4c13c294ea6b29c7e74de3951a004e8f3fd5b1c372bfbc3b5033"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.046585 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk9zf" event={"ID":"e5ba4e2d-a81f-4ab0-be7c-6ae18ba8f3cf","Type":"ContainerStarted","Data":"53e25d2e549513df5e3766210fdf09d1727841953f95fed9885d3cdb19913846"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.048779 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" event={"ID":"5a31ed34-05d3-4e0f-bb4b-03d086e01f84","Type":"ContainerStarted","Data":"dffe3f6d1050b13478b7e15ab25aca7b0af38a3e06b6487d1a1cd51d29c73415"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.053579 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" event={"ID":"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8","Type":"ContainerStarted","Data":"7c7fd9973de16e4423d230254b9ba870314175413f1a2911b5e17d6da927271e"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.058463 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv" event={"ID":"182965cc-4f7e-47e6-a33d-55aa53279d45","Type":"ContainerStarted","Data":"525563a56ac563f03ac8d6321c1af8154e0e0f1eb70705889f1a98bc5701f9be"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.061951 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8x7sr" event={"ID":"13d16a24-fc30-4d19-a3d5-1c55cc0021e4","Type":"ContainerStarted","Data":"7365ceed730f33f0cf52b9f4f09b9ac25201f03f46c97c60e2ace1e2ded5dd39"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.068706 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-h8pq2" event={"ID":"ff701efb-cabc-479a-ac05-4e9334e70385","Type":"ContainerStarted","Data":"f8c06102565677261f41465ca32a589ed4e8c7e28b20fa4235e1a2d72593ab7b"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.071006 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6vrqd" event={"ID":"5fb885c7-2a02-4adb-af59-35d8d9a1a3d9","Type":"ContainerStarted","Data":"5a10012125a2fb7cb9616c0254e2a1a272aa40ced19a62205328fe96e8cbfee2"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.076293 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c" event={"ID":"c5c791b9-d379-4311-b961-0927ca48702a","Type":"ContainerStarted","Data":"8b014b87ca2a9e635f27dc25c9798639c5547486ed420201dd6e573f46ff41e8"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.112072 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4nvsx" event={"ID":"e00acf3c-29bc-447d-bb80-7a1a8bf7cd84","Type":"ContainerStarted","Data":"2298251e0d56a60b71138ceb16af756ba1704846d9dc11758330028153b7ec35"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.112375 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2p58w" event={"ID":"61fba77d-c313-4c67-a0d8-7fbf2e172aa4","Type":"ContainerStarted","Data":"c3e43e30b485f5c105efb65e4fbc1f679cd48efc9b897cda50532a48d0f6390b"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.115598 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-h4pmt" event={"ID":"29721fdf-b040-414a-a6ca-6ba9dc544526","Type":"ContainerStarted","Data":"f08d737f363a14dca6949ff0028d0de06264a548a9a244b1aefb94898e4ce556"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.116645 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-h4pmt" Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.124128 4857 patch_prober.go:28] interesting pod/console-operator-58897d9998-h4pmt container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.124165 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-h4pmt" podUID="29721fdf-b040-414a-a6ca-6ba9dc544526" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.125520 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:57 crc kubenswrapper[4857]: E0222 00:08:57.125826 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:57.625817466 +0000 UTC m=+145.264546719 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.140248 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-f6z92" event={"ID":"816930b0-fec7-4eb4-bca1-32ca5775f5be","Type":"ContainerStarted","Data":"dbe7cf2100d790e2b0d2aa62adbe63156e2a87cd9dead22767616c8c992ad472"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.148249 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wq8b8" event={"ID":"f5acab23-9994-4c18-bc50-4f5ec7c7f525","Type":"ContainerStarted","Data":"ebf33c6d48d8fe0e0d14661defa1764b71efa1b6e6fb8bc87f653d7faa01d712"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.150491 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc" event={"ID":"ad13880d-4f52-4330-a02b-1f3006a9d979","Type":"ContainerStarted","Data":"21aea1830553b37cd5b59df922db3962ff190901023afbbc6a3a8d74b4f70258"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.160236 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-wc47x" event={"ID":"6ab42c49-5c7e-4bfb-9a4b-53e6eab21a33","Type":"ContainerStarted","Data":"8a627ce46675ba0da07f2173e12a316932669aeefe81e8cf713fb5b63fd6d542"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.178146 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4drbd" event={"ID":"e6ba7763-9bd5-45dd-a9a9-d47dc7f381ba","Type":"ContainerStarted","Data":"c3a03629d1c24a7bbd62d03b3ebceea0b66afadf6309a95ff0053fb61b3703f1"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.183849 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-m6mzr" event={"ID":"0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6","Type":"ContainerStarted","Data":"21931f876c693a2a31a1f7f9ba19d20cc5ddb7acb67d37cfb048392d2b561a06"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.187376 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" event={"ID":"05ee3f6d-2b50-4da4-8688-01ae8364d166","Type":"ContainerStarted","Data":"13a06d02575b7ac3689f040690d6ca2ca539e333164983aab2bb47c4982f5268"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.219786 4857 generic.go:334] "Generic (PLEG): container finished" podID="860d5a18-25c6-4484-bc45-2e741cd6b708" containerID="b43a85147007224099045fd1aa2c04fa4cbbdb5c7b9f0b831881447d4e76b787" exitCode=0 Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.219858 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-mx94d" event={"ID":"860d5a18-25c6-4484-bc45-2e741cd6b708","Type":"ContainerDied","Data":"b43a85147007224099045fd1aa2c04fa4cbbdb5c7b9f0b831881447d4e76b787"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.226291 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:57 crc kubenswrapper[4857]: E0222 00:08:57.228135 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:57.728116662 +0000 UTC m=+145.366845915 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.236189 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-cwlf7" event={"ID":"f9bce83c-ae40-482f-88d2-7de36a1a7ea5","Type":"ContainerStarted","Data":"54bbbcc321ada2e5f2b286124ea1f77db26c41b6f95b5ead5eae1c7b37a6ae66"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.236233 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-cwlf7" event={"ID":"f9bce83c-ae40-482f-88d2-7de36a1a7ea5","Type":"ContainerStarted","Data":"59dfb89250a11794fc51dd7572b3cc8b97b7577fde0a02c5dc984a9d04aea863"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.258359 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6d9wr" event={"ID":"0fa7f27c-4fa5-45db-b376-99c5c6468ec2","Type":"ContainerStarted","Data":"a3625d5f60b60c4fc98a0edca782769b04ba367b615bfb6463000bd9200350b3"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.268339 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29528640-g7sjv" event={"ID":"f07c5f6c-743d-4032-b70b-ecc0f514295d","Type":"ContainerStarted","Data":"129086f62452bf149b797fc37d111e6a4234c3e372415e4ccc6270e25e57f59b"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.276936 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4kmf" event={"ID":"e0b1777b-2a7f-4b10-8051-bb35cb18db72","Type":"ContainerStarted","Data":"c1e7af8e4033d482ee8273c727e6e1f42e278bddbff0a30d5eced5b825ef477e"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.286021 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-c4xt9" event={"ID":"e43cf3d5-a9c1-4804-a7cd-d0276c875daf","Type":"ContainerStarted","Data":"44b0e132d2ff4b60d5f7d8f819bacc68a5b6c68ddbd6d1d04641895ec0a31b34"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.300992 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv" event={"ID":"706cd825-7128-4833-8260-b3b73cccaac7","Type":"ContainerStarted","Data":"4d1c6f57af00f3a8960eac0264638d68b60ced27b28523fd19becf5a20b5f115"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.301438 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" podStartSLOduration=120.301409922 podStartE2EDuration="2m0.301409922s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:57.300807254 +0000 UTC m=+144.939536507" watchObservedRunningTime="2026-02-22 00:08:57.301409922 +0000 UTC m=+144.940139175" Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.314265 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-lfd8v" event={"ID":"1d0786e3-c21c-46f5-8ca9-7cad8ac62194","Type":"ContainerStarted","Data":"e66945f02d53ae62b767f23d5621ffe0055df8213a1eace8b908ad840c6095b9"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.324613 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nj6dw" event={"ID":"c2b40552-4007-4981-ba86-2d8864fc20cf","Type":"ContainerStarted","Data":"891b45392bdb60a2fac95ed893d5582f6d9ad14715b1ac2a79882f2bcecd4a78"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.327539 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:57 crc kubenswrapper[4857]: E0222 00:08:57.328892 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:57.828881055 +0000 UTC m=+145.467610308 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.343412 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-p9ztw" event={"ID":"d89916ae-f616-40ee-af76-fe17a677a488","Type":"ContainerStarted","Data":"d17cc2e7845009b07cb42a786cc2536712de1488903a25adef3ae47cd84bcc21"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.345664 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g4v7w" event={"ID":"614b8dc2-9b2a-4425-b0d0-78292729e035","Type":"ContainerStarted","Data":"a6e7faa59a4a0f064e55d4536a0b75c851639778a97636a94f9907a8c4dacab7"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.353164 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" event={"ID":"34787fe4-1fcc-4997-a612-f787f990c3d8","Type":"ContainerStarted","Data":"386e5a6bf559a795a091d1f0cef642945e955a0abf24122e2f59d50cafc6e9ed"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.384826 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g" event={"ID":"c719ea96-d4bf-4f14-b8bd-c32bb6a20c46","Type":"ContainerStarted","Data":"b080e3da656753e0731769e3616361b6585ed5714c3fd627ddc3f8e485dc45ad"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.405217 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9x24j" event={"ID":"7b02ac9a-c64c-41aa-b2ef-dbff783bf797","Type":"ContainerStarted","Data":"3828a74d0d1f7de622d84a2997b76075a7be41ed28e2ac675d6e3a3dad835902"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.417485 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" event={"ID":"fef422b7-19df-431d-b574-73457050f022","Type":"ContainerStarted","Data":"fd17ec2d5298f1cea103212c7f4187978e078a31df09b8bdf2eb27d3c181d9fb"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.424995 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-99cq6" event={"ID":"9e531e33-264a-498e-b279-cf687f1e6327","Type":"ContainerStarted","Data":"33cacefea17590c1ae6843b119f7eabcf5c73238554d6c21e4e87b9938f53324"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.430544 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:57 crc kubenswrapper[4857]: E0222 00:08:57.430787 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:57.93075979 +0000 UTC m=+145.569489033 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.431022 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:57 crc kubenswrapper[4857]: E0222 00:08:57.432117 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:57.932105337 +0000 UTC m=+145.570834680 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.438909 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" event={"ID":"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c","Type":"ContainerStarted","Data":"42a3dcf05f73f3c5133646209b8c96deec29c498ca560bf228dcf020f53d0cfe"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.447282 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6xx4f" event={"ID":"46391957-feff-4548-8c20-c713712585af","Type":"ContainerStarted","Data":"e0aaa06ea1a5dff1a2c4a2fb483e4cf577fc174ccfba80b4501ddf40c1059b6e"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.465318 4857 generic.go:334] "Generic (PLEG): container finished" podID="4edfa330-6a4a-4f86-b606-db0adb8301c8" containerID="45dd48d86348fcbdab7ad2ef89bd6a76d53e5e42bad527f38005970f3687326e" exitCode=0 Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.465486 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f" event={"ID":"4edfa330-6a4a-4f86-b606-db0adb8301c8","Type":"ContainerDied","Data":"45dd48d86348fcbdab7ad2ef89bd6a76d53e5e42bad527f38005970f3687326e"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.478015 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" event={"ID":"2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026","Type":"ContainerStarted","Data":"7412c6962c3041176435b9cc19f7bdc24c34b9b5cd51f1c505be61d32ab595a0"} Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.495678 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.506255 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.537646 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:57 crc kubenswrapper[4857]: E0222 00:08:57.537977 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:58.037955165 +0000 UTC m=+145.676684418 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.538388 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:57 crc kubenswrapper[4857]: E0222 00:08:57.539848 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:58.039834859 +0000 UTC m=+145.678564102 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.577173 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" podStartSLOduration=120.577156853 podStartE2EDuration="2m0.577156853s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:57.576114274 +0000 UTC m=+145.214843527" watchObservedRunningTime="2026-02-22 00:08:57.577156853 +0000 UTC m=+145.215886106" Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.643226 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:57 crc kubenswrapper[4857]: E0222 00:08:57.643517 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:58.143504015 +0000 UTC m=+145.782233268 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.744319 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:57 crc kubenswrapper[4857]: E0222 00:08:57.744961 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:58.244932996 +0000 UTC m=+145.883662299 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.796631 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-cwlf7" podStartSLOduration=120.79661579 podStartE2EDuration="2m0.79661579s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:57.79381572 +0000 UTC m=+145.432544973" watchObservedRunningTime="2026-02-22 00:08:57.79661579 +0000 UTC m=+145.435345043" Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.820435 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nj6dw" podStartSLOduration=121.820416318 podStartE2EDuration="2m1.820416318s" podCreationTimestamp="2026-02-22 00:06:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:57.819000578 +0000 UTC m=+145.457729831" watchObservedRunningTime="2026-02-22 00:08:57.820416318 +0000 UTC m=+145.459145571" Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.846611 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:57 crc kubenswrapper[4857]: E0222 00:08:57.847196 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:58.347176572 +0000 UTC m=+145.985905825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.874882 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-ctjrv" podStartSLOduration=120.87484863 podStartE2EDuration="2m0.87484863s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:57.8702595 +0000 UTC m=+145.508988773" watchObservedRunningTime="2026-02-22 00:08:57.87484863 +0000 UTC m=+145.513577883" Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.948597 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:57 crc kubenswrapper[4857]: E0222 00:08:57.949114 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:58.449101848 +0000 UTC m=+146.087831101 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.984731 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4nvsx" podStartSLOduration=120.984715723 podStartE2EDuration="2m0.984715723s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:57.938286879 +0000 UTC m=+145.577016132" watchObservedRunningTime="2026-02-22 00:08:57.984715723 +0000 UTC m=+145.623444976" Feb 22 00:08:57 crc kubenswrapper[4857]: I0222 00:08:57.999365 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.015252 4857 patch_prober.go:28] interesting pod/router-default-5444994796-lfd8v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 22 00:08:58 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Feb 22 00:08:58 crc kubenswrapper[4857]: [+]process-running ok Feb 22 00:08:58 crc kubenswrapper[4857]: healthz check failed Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.015297 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfd8v" podUID="1d0786e3-c21c-46f5-8ca9-7cad8ac62194" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.053421 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:58 crc kubenswrapper[4857]: E0222 00:08:58.053909 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:58.553893355 +0000 UTC m=+146.192622608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.063599 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-wq8b8" podStartSLOduration=121.063581782 podStartE2EDuration="2m1.063581782s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.058149847 +0000 UTC m=+145.696879100" watchObservedRunningTime="2026-02-22 00:08:58.063581782 +0000 UTC m=+145.702311035" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.133637 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-f6z92" podStartSLOduration=6.133612728 podStartE2EDuration="6.133612728s" podCreationTimestamp="2026-02-22 00:08:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.123220251 +0000 UTC m=+145.761949504" watchObservedRunningTime="2026-02-22 00:08:58.133612728 +0000 UTC m=+145.772341981" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.155134 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:58 crc kubenswrapper[4857]: E0222 00:08:58.155481 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:58.655467361 +0000 UTC m=+146.294196614 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.232427 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-l9f6g" podStartSLOduration=122.232408854 podStartE2EDuration="2m2.232408854s" podCreationTimestamp="2026-02-22 00:06:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.188511543 +0000 UTC m=+145.827240796" watchObservedRunningTime="2026-02-22 00:08:58.232408854 +0000 UTC m=+145.871138107" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.256694 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:58 crc kubenswrapper[4857]: E0222 00:08:58.257051 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:58.757022427 +0000 UTC m=+146.395751680 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.288564 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29528640-g7sjv" podStartSLOduration=122.288547565 podStartE2EDuration="2m2.288547565s" podCreationTimestamp="2026-02-22 00:06:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.287500366 +0000 UTC m=+145.926229619" watchObservedRunningTime="2026-02-22 00:08:58.288547565 +0000 UTC m=+145.927276818" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.358795 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:58 crc kubenswrapper[4857]: E0222 00:08:58.359091 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:58.859081296 +0000 UTC m=+146.497810549 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.381069 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-8sts2" podStartSLOduration=122.381051062 podStartE2EDuration="2m2.381051062s" podCreationTimestamp="2026-02-22 00:06:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.321241988 +0000 UTC m=+145.959971241" watchObservedRunningTime="2026-02-22 00:08:58.381051062 +0000 UTC m=+146.019780315" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.424880 4857 csr.go:261] certificate signing request csr-4k5mx is approved, waiting to be issued Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.447794 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6d9wr" podStartSLOduration=121.447776924 podStartE2EDuration="2m1.447776924s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.368414792 +0000 UTC m=+146.007144045" watchObservedRunningTime="2026-02-22 00:08:58.447776924 +0000 UTC m=+146.086506177" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.448351 4857 csr.go:257] certificate signing request csr-4k5mx is issued Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.461395 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:58 crc kubenswrapper[4857]: E0222 00:08:58.461729 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:58.961714482 +0000 UTC m=+146.600443735 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.479603 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6vrqd" podStartSLOduration=121.479568011 podStartE2EDuration="2m1.479568011s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.477328148 +0000 UTC m=+146.116057401" watchObservedRunningTime="2026-02-22 00:08:58.479568011 +0000 UTC m=+146.118297264" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.538456 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-h4pmt" podStartSLOduration=121.53844093 podStartE2EDuration="2m1.53844093s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.537585815 +0000 UTC m=+146.176315058" watchObservedRunningTime="2026-02-22 00:08:58.53844093 +0000 UTC m=+146.177170183" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.539532 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2p58w" event={"ID":"61fba77d-c313-4c67-a0d8-7fbf2e172aa4","Type":"ContainerStarted","Data":"3ac856e3ee9dc3f5ca0570e9ca79d0328b8f689a84cf02603664477a34ed79cb"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.540280 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2p58w" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.562311 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-c4xt9" event={"ID":"e43cf3d5-a9c1-4804-a7cd-d0276c875daf","Type":"ContainerStarted","Data":"bebeccbc699fb2c88ac1c765d430caa53427a844bef3230b0e1558b2b2d435ea"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.568815 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:58 crc kubenswrapper[4857]: E0222 00:08:58.569263 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:59.069246888 +0000 UTC m=+146.707976131 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.581817 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk9zf" event={"ID":"e5ba4e2d-a81f-4ab0-be7c-6ae18ba8f3cf","Type":"ContainerStarted","Data":"cc3ad385bbbf0ff5b11560b9c24ea35c1b5b6343f4ec17fee7627b8785d80fb5"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.581863 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk9zf" event={"ID":"e5ba4e2d-a81f-4ab0-be7c-6ae18ba8f3cf","Type":"ContainerStarted","Data":"0b53e862d2a9887b75329ffa6814c6d3beb8ab54109229e85d8e7be5d2b00c3d"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.599854 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4kmf" event={"ID":"e0b1777b-2a7f-4b10-8051-bb35cb18db72","Type":"ContainerStarted","Data":"0c84f43295e7801a9bf80443f95927b4c812cb7b534f0c7409bfe252ed92670e"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.611880 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" podStartSLOduration=121.611862213 podStartE2EDuration="2m1.611862213s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.568172737 +0000 UTC m=+146.206901990" watchObservedRunningTime="2026-02-22 00:08:58.611862213 +0000 UTC m=+146.250591466" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.616504 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" event={"ID":"56f6f504-2f13-4ae2-b5cd-d67ec7c140a7","Type":"ContainerStarted","Data":"de70b1af07224118cdfb3df3c224df5b7cb551b1a5eec373a006a4d6cb11caf7"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.627479 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-wc47x" event={"ID":"6ab42c49-5c7e-4bfb-9a4b-53e6eab21a33","Type":"ContainerStarted","Data":"bf0899b1326b4ff9b5af4a92512827caefc1d499bb9393f7eec42489558070f4"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.635795 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-lfd8v" podStartSLOduration=121.635780815 podStartE2EDuration="2m1.635780815s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.630740301 +0000 UTC m=+146.269469554" watchObservedRunningTime="2026-02-22 00:08:58.635780815 +0000 UTC m=+146.274510068" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.636343 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4drbd" podStartSLOduration=121.636336831 podStartE2EDuration="2m1.636336831s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.612127521 +0000 UTC m=+146.250856774" watchObservedRunningTime="2026-02-22 00:08:58.636336831 +0000 UTC m=+146.275066084" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.639097 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g4v7w" event={"ID":"614b8dc2-9b2a-4425-b0d0-78292729e035","Type":"ContainerStarted","Data":"18922540a6eecb509dd9529462b75ad48699b8756e226d06d8f0c93ef5da9280"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.657971 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" event={"ID":"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c","Type":"ContainerStarted","Data":"781a50a624c5adcbde873dbe842de412ec8284a660bd60e0a29009b8a5282985"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.658523 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.663596 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-d4kmf" podStartSLOduration=121.663577777 podStartE2EDuration="2m1.663577777s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.661341003 +0000 UTC m=+146.300070246" watchObservedRunningTime="2026-02-22 00:08:58.663577777 +0000 UTC m=+146.302307030" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.676562 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:58 crc kubenswrapper[4857]: E0222 00:08:58.679406 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:59.179385118 +0000 UTC m=+146.818114371 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.683184 4857 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-lm22m container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.30:6443/healthz\": dial tcp 10.217.0.30:6443: connect: connection refused" start-of-body= Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.683245 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" podUID="b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.30:6443/healthz\": dial tcp 10.217.0.30:6443: connect: connection refused" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.684239 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6xx4f" event={"ID":"46391957-feff-4548-8c20-c713712585af","Type":"ContainerStarted","Data":"8f8f581642708651d5ee17544ca9b9df2a014da4d56c1865de9e245189a45e5e"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.684269 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6xx4f" event={"ID":"46391957-feff-4548-8c20-c713712585af","Type":"ContainerStarted","Data":"78fc6bdaa8dfc4bc55d8d7a3ba457f2c13d740955af7aa385493549b788be9cd"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.713677 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-zk9zf" podStartSLOduration=122.713660695 podStartE2EDuration="2m2.713660695s" podCreationTimestamp="2026-02-22 00:06:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.710888056 +0000 UTC m=+146.349617309" watchObservedRunningTime="2026-02-22 00:08:58.713660695 +0000 UTC m=+146.352389948" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.720245 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tctk6" event={"ID":"e8865dea-cf78-4acd-bcb4-aad234172e99","Type":"ContainerStarted","Data":"8dfbec931054be6ada2ccf2b31ea298eafd9add2531bc7191aeb166690db3e26"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.759414 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc" event={"ID":"ad13880d-4f52-4330-a02b-1f3006a9d979","Type":"ContainerStarted","Data":"a25a7649042f24831d987f4fd504e295d028ad3a673939c55e73b4c51797f71a"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.761003 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2p58w" podStartSLOduration=121.760984814 podStartE2EDuration="2m1.760984814s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.75839355 +0000 UTC m=+146.397122803" watchObservedRunningTime="2026-02-22 00:08:58.760984814 +0000 UTC m=+146.399714067" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.783063 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:58 crc kubenswrapper[4857]: E0222 00:08:58.785189 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:59.285177084 +0000 UTC m=+146.923906337 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.791309 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8x7sr" event={"ID":"13d16a24-fc30-4d19-a3d5-1c55cc0021e4","Type":"ContainerStarted","Data":"92f6dfab5ef0b1ee70f8801573f800aace9f1ab731fba0f364d860a2d14076ec"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.791347 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8x7sr" event={"ID":"13d16a24-fc30-4d19-a3d5-1c55cc0021e4","Type":"ContainerStarted","Data":"ab423b9a48d61296d0b977129ee9074a9cea021ddb76327feb482965530c780c"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.799908 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g4v7w" podStartSLOduration=121.799891614 podStartE2EDuration="2m1.799891614s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.799602606 +0000 UTC m=+146.438331859" watchObservedRunningTime="2026-02-22 00:08:58.799891614 +0000 UTC m=+146.438620887" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.831457 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" podStartSLOduration=122.831439983 podStartE2EDuration="2m2.831439983s" podCreationTimestamp="2026-02-22 00:06:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.828597432 +0000 UTC m=+146.467326685" watchObservedRunningTime="2026-02-22 00:08:58.831439983 +0000 UTC m=+146.470169236" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.858413 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" event={"ID":"34787fe4-1fcc-4997-a612-f787f990c3d8","Type":"ContainerStarted","Data":"5d8726d032d8916ec6af882cf6225d743e5ddd038f1d2042d974e16c3d7ea6d9"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.860379 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.864571 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-h8pq2" event={"ID":"ff701efb-cabc-479a-ac05-4e9334e70385","Type":"ContainerStarted","Data":"194fe7ec56e810d6139e31bf87a6616ecaddf2dc9fa90600c795438661dac048"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.865557 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-h8pq2" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.879424 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv" event={"ID":"706cd825-7128-4833-8260-b3b73cccaac7","Type":"ContainerStarted","Data":"f148ea88b5663ece2619af485f301e73d17442bd36377d16d011847cb0e2af20"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.881548 4857 patch_prober.go:28] interesting pod/downloads-7954f5f757-h8pq2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.881584 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-h8pq2" podUID="ff701efb-cabc-479a-ac05-4e9334e70385" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.883468 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c" event={"ID":"c5c791b9-d379-4311-b961-0927ca48702a","Type":"ContainerStarted","Data":"d2ed44b943d8879aca0f85da540995378eeed2deaccdbb4a62a85e5197e5bd3f"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.884290 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.884886 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:58 crc kubenswrapper[4857]: E0222 00:08:58.885115 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:59.385095123 +0000 UTC m=+147.023824376 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.885340 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:58 crc kubenswrapper[4857]: E0222 00:08:58.886551 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:59.386536484 +0000 UTC m=+147.025265737 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.889152 4857 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tm4ld container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:5443/healthz\": dial tcp 10.217.0.18:5443: connect: connection refused" start-of-body= Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.889194 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" podUID="34787fe4-1fcc-4997-a612-f787f990c3d8" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.18:5443/healthz\": dial tcp 10.217.0.18:5443: connect: connection refused" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.899253 4857 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-7wd8c container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.899312 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c" podUID="c5c791b9-d379-4311-b961-0927ca48702a" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.899239 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tctk6" podStartSLOduration=121.899222355 podStartE2EDuration="2m1.899222355s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.860654426 +0000 UTC m=+146.499383689" watchObservedRunningTime="2026-02-22 00:08:58.899222355 +0000 UTC m=+146.537951608" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.899790 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-p9ztw" event={"ID":"d89916ae-f616-40ee-af76-fe17a677a488","Type":"ContainerStarted","Data":"87ead1d4a7ff626a8dd2791758eaeb493e242c77031103248444a41cc6029378"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.900234 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-6xx4f" podStartSLOduration=121.900223924 podStartE2EDuration="2m1.900223924s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.899429661 +0000 UTC m=+146.538158914" watchObservedRunningTime="2026-02-22 00:08:58.900223924 +0000 UTC m=+146.538953177" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.942879 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv" event={"ID":"182965cc-4f7e-47e6-a33d-55aa53279d45","Type":"ContainerStarted","Data":"ff1ddda8244a4306923f01ccce647f84608963a9866624baff7d6042a4c33be1"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.943244 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv" event={"ID":"182965cc-4f7e-47e6-a33d-55aa53279d45","Type":"ContainerStarted","Data":"deb2c69a6d4eba5e822806fa84e03a57150de5add0fa6dd49eb08845e9f5b729"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.957421 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-spwkv" podStartSLOduration=121.957405354 podStartE2EDuration="2m1.957405354s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:58.957114776 +0000 UTC m=+146.595844049" watchObservedRunningTime="2026-02-22 00:08:58.957405354 +0000 UTC m=+146.596134607" Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.973708 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-4drbd" event={"ID":"e6ba7763-9bd5-45dd-a9a9-d47dc7f381ba","Type":"ContainerStarted","Data":"b25518e41427f52f5f20e11d2860165700352130af5ef73be56d2c75a156f297"} Feb 22 00:08:58 crc kubenswrapper[4857]: I0222 00:08:58.990112 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:58 crc kubenswrapper[4857]: E0222 00:08:58.991587 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:59.491568818 +0000 UTC m=+147.130298071 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.007974 4857 patch_prober.go:28] interesting pod/router-default-5444994796-lfd8v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 22 00:08:59 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Feb 22 00:08:59 crc kubenswrapper[4857]: [+]process-running ok Feb 22 00:08:59 crc kubenswrapper[4857]: healthz check failed Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.008892 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" event={"ID":"05ee3f6d-2b50-4da4-8688-01ae8364d166","Type":"ContainerStarted","Data":"05e9778d99c15d2c5cbae7799ef80b8be54fdda4bbc44e1549eae4965a464f1c"} Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.021134 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfd8v" podUID="1d0786e3-c21c-46f5-8ca9-7cad8ac62194" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.034007 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-8x7sr" podStartSLOduration=122.033985888 podStartE2EDuration="2m2.033985888s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:59.0333654 +0000 UTC m=+146.672094663" watchObservedRunningTime="2026-02-22 00:08:59.033985888 +0000 UTC m=+146.672715141" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.039326 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-99cq6" event={"ID":"9e531e33-264a-498e-b279-cf687f1e6327","Type":"ContainerStarted","Data":"812215d00ad1cacd15db4d5dc29b4414a3901c537087c2d4cd78702bf93cd717"} Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.074518 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c" podStartSLOduration=122.074501013 podStartE2EDuration="2m2.074501013s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:59.073868665 +0000 UTC m=+146.712597918" watchObservedRunningTime="2026-02-22 00:08:59.074501013 +0000 UTC m=+146.713230266" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.088620 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" event={"ID":"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8","Type":"ContainerStarted","Data":"1837b66f71bf7805f8b4307a67a79a46079d745062c0638df7ba39b0091df73b"} Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.091926 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:59 crc kubenswrapper[4857]: E0222 00:08:59.095257 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:59.595243434 +0000 UTC m=+147.233972687 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.130204 4857 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vmk2g container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.130263 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" podUID="4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.132990 4857 patch_prober.go:28] interesting pod/console-operator-58897d9998-h4pmt container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.133156 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-h4pmt" podUID="29721fdf-b040-414a-a6ca-6ba9dc544526" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.136686 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-m6mzr" event={"ID":"0f0bd158-5aa1-4c9a-8b64-14aa0a6343c6","Type":"ContainerStarted","Data":"8a13448a4c9e0f825403f5f53d4d0340352b3a078174476b6a13c5e4520fb8d5"} Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.138241 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.138284 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.138296 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" event={"ID":"2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026","Type":"ContainerStarted","Data":"8f342c0a9862fc8bf02e7c247a2a46dfa620df86ab18b7ba8f8507985e7e572c"} Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.157350 4857 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-nqshz container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" start-of-body= Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.157404 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" podUID="2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.175731 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" podStartSLOduration=122.175713068 podStartE2EDuration="2m2.175713068s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:59.133612138 +0000 UTC m=+146.772341391" watchObservedRunningTime="2026-02-22 00:08:59.175713068 +0000 UTC m=+146.814442321" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.176895 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hm6qc" podStartSLOduration=122.176889772 podStartE2EDuration="2m2.176889772s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:59.173918037 +0000 UTC m=+146.812647290" watchObservedRunningTime="2026-02-22 00:08:59.176889772 +0000 UTC m=+146.815619025" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.195594 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:59 crc kubenswrapper[4857]: E0222 00:08:59.197422 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:59.697399787 +0000 UTC m=+147.336129040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.201099 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-h8pq2" podStartSLOduration=122.201084882 podStartE2EDuration="2m2.201084882s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:59.198024534 +0000 UTC m=+146.836753777" watchObservedRunningTime="2026-02-22 00:08:59.201084882 +0000 UTC m=+146.839814135" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.274157 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.274887 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.277110 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-p9ztw" podStartSLOduration=7.277090699 podStartE2EDuration="7.277090699s" podCreationTimestamp="2026-02-22 00:08:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:59.273327001 +0000 UTC m=+146.912056254" watchObservedRunningTime="2026-02-22 00:08:59.277090699 +0000 UTC m=+146.915819962" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.298295 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:59 crc kubenswrapper[4857]: E0222 00:08:59.298638 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:08:59.798626702 +0000 UTC m=+147.437355945 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.351532 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" podStartSLOduration=122.351513711 podStartE2EDuration="2m2.351513711s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:59.351287264 +0000 UTC m=+146.990016518" watchObservedRunningTime="2026-02-22 00:08:59.351513711 +0000 UTC m=+146.990242964" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.400217 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:59 crc kubenswrapper[4857]: E0222 00:08:59.400560 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:08:59.900545699 +0000 UTC m=+147.539274952 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.432112 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-m6mzr" podStartSLOduration=122.432095358 podStartE2EDuration="2m2.432095358s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:59.43147018 +0000 UTC m=+147.070199433" watchObservedRunningTime="2026-02-22 00:08:59.432095358 +0000 UTC m=+147.070824611" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.441622 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-99cq6" podStartSLOduration=122.441599289 podStartE2EDuration="2m2.441599289s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:59.389298388 +0000 UTC m=+147.028027661" watchObservedRunningTime="2026-02-22 00:08:59.441599289 +0000 UTC m=+147.080328542" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.451143 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-22 00:03:58 +0000 UTC, rotation deadline is 2026-12-19 17:57:25.383199921 +0000 UTC Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.451178 4857 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7217h48m25.932024749s for next certificate rotation Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.460236 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" podStartSLOduration=122.4602234 podStartE2EDuration="2m2.4602234s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:59.457237185 +0000 UTC m=+147.095966438" watchObservedRunningTime="2026-02-22 00:08:59.4602234 +0000 UTC m=+147.098952653" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.492225 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xwszv" podStartSLOduration=122.492208692 podStartE2EDuration="2m2.492208692s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:59.490316408 +0000 UTC m=+147.129045661" watchObservedRunningTime="2026-02-22 00:08:59.492208692 +0000 UTC m=+147.130937945" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.503741 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:59 crc kubenswrapper[4857]: E0222 00:08:59.504101 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.004089411 +0000 UTC m=+147.642818664 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.509578 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" podStartSLOduration=123.509560947 podStartE2EDuration="2m3.509560947s" podCreationTimestamp="2026-02-22 00:06:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:08:59.508235739 +0000 UTC m=+147.146964982" watchObservedRunningTime="2026-02-22 00:08:59.509560947 +0000 UTC m=+147.148290200" Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.605485 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:59 crc kubenswrapper[4857]: E0222 00:08:59.605723 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.105693338 +0000 UTC m=+147.744422591 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.605958 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:59 crc kubenswrapper[4857]: E0222 00:08:59.606305 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.106293545 +0000 UTC m=+147.745022798 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.707435 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:59 crc kubenswrapper[4857]: E0222 00:08:59.707642 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.207611143 +0000 UTC m=+147.846340396 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.809137 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:59 crc kubenswrapper[4857]: E0222 00:08:59.809502 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.309485918 +0000 UTC m=+147.948215171 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.910534 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:08:59 crc kubenswrapper[4857]: E0222 00:08:59.910701 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.410658552 +0000 UTC m=+148.049387805 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:08:59 crc kubenswrapper[4857]: I0222 00:08:59.910991 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:08:59 crc kubenswrapper[4857]: E0222 00:08:59.911369 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.411352642 +0000 UTC m=+148.050081935 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.003236 4857 patch_prober.go:28] interesting pod/router-default-5444994796-lfd8v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 22 00:09:00 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Feb 22 00:09:00 crc kubenswrapper[4857]: [+]process-running ok Feb 22 00:09:00 crc kubenswrapper[4857]: healthz check failed Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.003297 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfd8v" podUID="1d0786e3-c21c-46f5-8ca9-7cad8ac62194" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.012062 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:09:00 crc kubenswrapper[4857]: E0222 00:09:00.012234 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.512209438 +0000 UTC m=+148.150938691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.012525 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:00 crc kubenswrapper[4857]: E0222 00:09:00.012860 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.512849265 +0000 UTC m=+148.151578518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.113468 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:09:00 crc kubenswrapper[4857]: E0222 00:09:00.113652 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.613617758 +0000 UTC m=+148.252347011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.113756 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:00 crc kubenswrapper[4857]: E0222 00:09:00.114048 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.61402278 +0000 UTC m=+148.252752023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.139062 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-c4xt9" event={"ID":"e43cf3d5-a9c1-4804-a7cd-d0276c875daf","Type":"ContainerStarted","Data":"952f09d324d11c5cef6bd7b155938f4caa92a5fe6205d689b550b7682c7eb5a4"} Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.139139 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-c4xt9" Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.140740 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f" event={"ID":"4edfa330-6a4a-4f86-b606-db0adb8301c8","Type":"ContainerStarted","Data":"b435e3e8fea419ebac739f882fdedb68e79444b7596d0be64e91a6a331417e23"} Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.140875 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f" Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.142083 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tctk6" event={"ID":"e8865dea-cf78-4acd-bcb4-aad234172e99","Type":"ContainerStarted","Data":"3fc4fc626efacbdb6dd29d47de607192080f081c8ffaee517de94690b93f5d39"} Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.144186 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-mx94d" event={"ID":"860d5a18-25c6-4484-bc45-2e741cd6b708","Type":"ContainerStarted","Data":"16185469ea4e42ae23d16320e47b4925a1ba553a1558dd4a6d1f67e9105ddb57"} Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.144228 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-mx94d" event={"ID":"860d5a18-25c6-4484-bc45-2e741cd6b708","Type":"ContainerStarted","Data":"b4904038c5047ac761df1793e72a80fb14ff15d9e02d31991f955bf40a6964ce"} Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.148624 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9x24j" event={"ID":"7b02ac9a-c64c-41aa-b2ef-dbff783bf797","Type":"ContainerStarted","Data":"59515bacb9c68fe09c18128e62089fa26955733f62ba7d0a34f25b336874d9dc"} Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.151588 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-wc47x" event={"ID":"6ab42c49-5c7e-4bfb-9a4b-53e6eab21a33","Type":"ContainerStarted","Data":"b3d4e102c5918397c4986bb1f81b1f5f1e1c56c8ed74a7fe59464b3c5b652659"} Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.151668 4857 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-nqshz container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" start-of-body= Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.151737 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" podUID="2f5ae264-3cdf-4fc2-8fb0-f25bcc98f026" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.151979 4857 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-tm4ld container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:5443/healthz\": dial tcp 10.217.0.18:5443: connect: connection refused" start-of-body= Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.152025 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" podUID="34787fe4-1fcc-4997-a612-f787f990c3d8" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.18:5443/healthz\": dial tcp 10.217.0.18:5443: connect: connection refused" Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.153428 4857 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-lm22m container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.30:6443/healthz\": dial tcp 10.217.0.30:6443: connect: connection refused" start-of-body= Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.153491 4857 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vmk2g container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.153562 4857 patch_prober.go:28] interesting pod/downloads-7954f5f757-h8pq2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.153588 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" podUID="4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.153609 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-h8pq2" podUID="ff701efb-cabc-479a-ac05-4e9334e70385" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.153534 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" podUID="b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.30:6443/healthz\": dial tcp 10.217.0.30:6443: connect: connection refused" Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.207672 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-c4xt9" podStartSLOduration=8.20765565 podStartE2EDuration="8.20765565s" podCreationTimestamp="2026-02-22 00:08:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:09:00.205463478 +0000 UTC m=+147.844192731" watchObservedRunningTime="2026-02-22 00:09:00.20765565 +0000 UTC m=+147.846384903" Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.211540 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-7wd8c" Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.215613 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:09:00 crc kubenswrapper[4857]: E0222 00:09:00.215770 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.715739 +0000 UTC m=+148.354468253 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.216764 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:00 crc kubenswrapper[4857]: E0222 00:09:00.219114 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.719099576 +0000 UTC m=+148.357828829 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.271573 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-mx94d" podStartSLOduration=124.271554401 podStartE2EDuration="2m4.271554401s" podCreationTimestamp="2026-02-22 00:06:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:09:00.238050626 +0000 UTC m=+147.876779879" watchObservedRunningTime="2026-02-22 00:09:00.271554401 +0000 UTC m=+147.910283654" Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.272539 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-wc47x" podStartSLOduration=123.27253203 podStartE2EDuration="2m3.27253203s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:09:00.269982756 +0000 UTC m=+147.908712009" watchObservedRunningTime="2026-02-22 00:09:00.27253203 +0000 UTC m=+147.911261283" Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.298026 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.318285 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:09:00 crc kubenswrapper[4857]: E0222 00:09:00.318698 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.818680845 +0000 UTC m=+148.457410098 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.319722 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f" podStartSLOduration=124.319692444 podStartE2EDuration="2m4.319692444s" podCreationTimestamp="2026-02-22 00:06:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:09:00.317716557 +0000 UTC m=+147.956445810" watchObservedRunningTime="2026-02-22 00:09:00.319692444 +0000 UTC m=+147.958421697" Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.420226 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:00 crc kubenswrapper[4857]: E0222 00:09:00.420617 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:00.920597991 +0000 UTC m=+148.559327234 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.521241 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:09:00 crc kubenswrapper[4857]: E0222 00:09:00.521415 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:01.021388235 +0000 UTC m=+148.660117488 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.522058 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:00 crc kubenswrapper[4857]: E0222 00:09:00.522402 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:01.022392613 +0000 UTC m=+148.661121866 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.622910 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:09:00 crc kubenswrapper[4857]: E0222 00:09:00.623111 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:01.123085673 +0000 UTC m=+148.761814926 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.623373 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:00 crc kubenswrapper[4857]: E0222 00:09:00.623639 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:01.123631259 +0000 UTC m=+148.762360512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.724971 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:09:00 crc kubenswrapper[4857]: E0222 00:09:00.725125 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:01.225100032 +0000 UTC m=+148.863829285 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.725245 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:00 crc kubenswrapper[4857]: E0222 00:09:00.725588 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:01.225572956 +0000 UTC m=+148.864302209 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.826377 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:09:00 crc kubenswrapper[4857]: E0222 00:09:00.826575 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:01.326549285 +0000 UTC m=+148.965278538 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.827029 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:00 crc kubenswrapper[4857]: E0222 00:09:00.827363 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:01.327351527 +0000 UTC m=+148.966080850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.884554 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-h4pmt" Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.928715 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:09:00 crc kubenswrapper[4857]: E0222 00:09:00.928897 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:01.428873222 +0000 UTC m=+149.067602475 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.928981 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.930131 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.930963 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:09:00 crc kubenswrapper[4857]: I0222 00:09:00.949979 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.004288 4857 patch_prober.go:28] interesting pod/router-default-5444994796-lfd8v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 22 00:09:01 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Feb 22 00:09:01 crc kubenswrapper[4857]: [+]process-running ok Feb 22 00:09:01 crc kubenswrapper[4857]: healthz check failed Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.004389 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfd8v" podUID="1d0786e3-c21c-46f5-8ca9-7cad8ac62194" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.031187 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.031233 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.031309 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:09:01 crc kubenswrapper[4857]: E0222 00:09:01.031831 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:01.531816186 +0000 UTC m=+149.170545439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.034854 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.035903 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.098147 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.108337 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.131987 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:09:01 crc kubenswrapper[4857]: E0222 00:09:01.132278 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:01.632183659 +0000 UTC m=+149.270912912 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.132514 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:01 crc kubenswrapper[4857]: E0222 00:09:01.132901 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:01.632890898 +0000 UTC m=+149.271620151 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.160485 4857 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vmk2g container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" start-of-body= Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.160539 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" podUID="4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": dial tcp 10.217.0.33:8080: connect: connection refused" Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.160651 4857 patch_prober.go:28] interesting pod/downloads-7954f5f757-h8pq2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.160672 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-h8pq2" podUID="ff701efb-cabc-479a-ac05-4e9334e70385" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.169286 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-9nx68" Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.172676 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.187383 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-tm4ld" Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.193016 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.234604 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:09:01 crc kubenswrapper[4857]: E0222 00:09:01.235462 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:01.735436552 +0000 UTC m=+149.374165805 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.239027 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:01 crc kubenswrapper[4857]: E0222 00:09:01.282965 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:01.782951807 +0000 UTC m=+149.421681060 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.340850 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:09:01 crc kubenswrapper[4857]: E0222 00:09:01.341312 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:01.8412969 +0000 UTC m=+149.480026143 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.445444 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:01 crc kubenswrapper[4857]: E0222 00:09:01.446106 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:01.946091798 +0000 UTC m=+149.584821051 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.546735 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:09:01 crc kubenswrapper[4857]: E0222 00:09:01.547166 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:02.047118108 +0000 UTC m=+149.685847361 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.649396 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:01 crc kubenswrapper[4857]: E0222 00:09:01.649974 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:02.149962451 +0000 UTC m=+149.788691704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.749859 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:09:01 crc kubenswrapper[4857]: E0222 00:09:01.749947 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:02.24992726 +0000 UTC m=+149.888656523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.750483 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:01 crc kubenswrapper[4857]: E0222 00:09:01.751020 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:02.251010841 +0000 UTC m=+149.889740094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.817211 4857 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.851964 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:09:01 crc kubenswrapper[4857]: E0222 00:09:01.852431 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:02.352411762 +0000 UTC m=+149.991141015 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.854265 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4dw8f" Feb 22 00:09:01 crc kubenswrapper[4857]: I0222 00:09:01.953407 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:01 crc kubenswrapper[4857]: E0222 00:09:01.953869 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:02.453850504 +0000 UTC m=+150.092579757 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.007786 4857 patch_prober.go:28] interesting pod/router-default-5444994796-lfd8v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 22 00:09:02 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Feb 22 00:09:02 crc kubenswrapper[4857]: [+]process-running ok Feb 22 00:09:02 crc kubenswrapper[4857]: healthz check failed Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.007858 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfd8v" podUID="1d0786e3-c21c-46f5-8ca9-7cad8ac62194" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.054824 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:09:02 crc kubenswrapper[4857]: E0222 00:09:02.054991 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:02.554967497 +0000 UTC m=+150.193696750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.055191 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:02 crc kubenswrapper[4857]: E0222 00:09:02.055501 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:02.555485662 +0000 UTC m=+150.194214965 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.085834 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5rpdr"] Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.086769 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rpdr" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.098620 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.101905 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5rpdr"] Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.156194 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:09:02 crc kubenswrapper[4857]: E0222 00:09:02.156409 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-22 00:09:02.656359568 +0000 UTC m=+150.295088821 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.156506 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa45663f-d64c-431c-be1d-29175cc804ea-catalog-content\") pod \"community-operators-5rpdr\" (UID: \"fa45663f-d64c-431c-be1d-29175cc804ea\") " pod="openshift-marketplace/community-operators-5rpdr" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.156536 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhpbm\" (UniqueName: \"kubernetes.io/projected/fa45663f-d64c-431c-be1d-29175cc804ea-kube-api-access-mhpbm\") pod \"community-operators-5rpdr\" (UID: \"fa45663f-d64c-431c-be1d-29175cc804ea\") " pod="openshift-marketplace/community-operators-5rpdr" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.156581 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.156604 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa45663f-d64c-431c-be1d-29175cc804ea-utilities\") pod \"community-operators-5rpdr\" (UID: \"fa45663f-d64c-431c-be1d-29175cc804ea\") " pod="openshift-marketplace/community-operators-5rpdr" Feb 22 00:09:02 crc kubenswrapper[4857]: E0222 00:09:02.156937 4857 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-22 00:09:02.656921433 +0000 UTC m=+150.295650686 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-96hns" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.165343 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"0a657bc057c6f9c908544e0ce27878eba6c63f315ce0a862399f3d65720880e2"} Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.165426 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"5c15ae515db90f5a11ef1e3291371563b3ddbe2fe58b4e7b7d37eab0a4885852"} Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.170425 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"16d2172e48904d3f0b8588363c023ebc5841041c3e8e649d2da6367cdf4911e1"} Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.170459 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"01b3fdbd8d9f94269d60e59a56db63c8f706a961a8fecded72eda695688e1ae6"} Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.170856 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.173101 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9x24j" event={"ID":"7b02ac9a-c64c-41aa-b2ef-dbff783bf797","Type":"ContainerStarted","Data":"4203a004705073a2bc624ff4b797fc45637ef06824d94a58784011079e5b72ab"} Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.173146 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9x24j" event={"ID":"7b02ac9a-c64c-41aa-b2ef-dbff783bf797","Type":"ContainerStarted","Data":"93d061b18e3a27e0f112445f3b3334529761f7e2ce8c340e0fa7f55ca8bd1ad0"} Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.175817 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"cfdd7c8c72042189148493560ff5f6671041b190c7e9bda3ee9db2e213f60076"} Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.175888 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"fd5d53ebb221df756a6ded39a04fd99ef7be0fb29e1d67d65b47573c83339ccd"} Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.225249 4857 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-22T00:09:01.81725141Z","Handler":null,"Name":""} Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.227589 4857 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.227621 4857 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.257713 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.258167 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa45663f-d64c-431c-be1d-29175cc804ea-catalog-content\") pod \"community-operators-5rpdr\" (UID: \"fa45663f-d64c-431c-be1d-29175cc804ea\") " pod="openshift-marketplace/community-operators-5rpdr" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.258196 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhpbm\" (UniqueName: \"kubernetes.io/projected/fa45663f-d64c-431c-be1d-29175cc804ea-kube-api-access-mhpbm\") pod \"community-operators-5rpdr\" (UID: \"fa45663f-d64c-431c-be1d-29175cc804ea\") " pod="openshift-marketplace/community-operators-5rpdr" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.258359 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa45663f-d64c-431c-be1d-29175cc804ea-utilities\") pod \"community-operators-5rpdr\" (UID: \"fa45663f-d64c-431c-be1d-29175cc804ea\") " pod="openshift-marketplace/community-operators-5rpdr" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.259427 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa45663f-d64c-431c-be1d-29175cc804ea-catalog-content\") pod \"community-operators-5rpdr\" (UID: \"fa45663f-d64c-431c-be1d-29175cc804ea\") " pod="openshift-marketplace/community-operators-5rpdr" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.259721 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa45663f-d64c-431c-be1d-29175cc804ea-utilities\") pod \"community-operators-5rpdr\" (UID: \"fa45663f-d64c-431c-be1d-29175cc804ea\") " pod="openshift-marketplace/community-operators-5rpdr" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.271127 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.280537 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m5rft"] Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.281412 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5rft" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.284481 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.288764 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhpbm\" (UniqueName: \"kubernetes.io/projected/fa45663f-d64c-431c-be1d-29175cc804ea-kube-api-access-mhpbm\") pod \"community-operators-5rpdr\" (UID: \"fa45663f-d64c-431c-be1d-29175cc804ea\") " pod="openshift-marketplace/community-operators-5rpdr" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.315717 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m5rft"] Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.359340 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18611d84-22b3-49df-9569-ac383eb42cab-catalog-content\") pod \"certified-operators-m5rft\" (UID: \"18611d84-22b3-49df-9569-ac383eb42cab\") " pod="openshift-marketplace/certified-operators-m5rft" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.359396 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.359434 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18611d84-22b3-49df-9569-ac383eb42cab-utilities\") pod \"certified-operators-m5rft\" (UID: \"18611d84-22b3-49df-9569-ac383eb42cab\") " pod="openshift-marketplace/certified-operators-m5rft" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.359484 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59kjl\" (UniqueName: \"kubernetes.io/projected/18611d84-22b3-49df-9569-ac383eb42cab-kube-api-access-59kjl\") pod \"certified-operators-m5rft\" (UID: \"18611d84-22b3-49df-9569-ac383eb42cab\") " pod="openshift-marketplace/certified-operators-m5rft" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.362080 4857 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.362135 4857 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.385850 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-96hns\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.392953 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.401672 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rpdr" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.460247 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18611d84-22b3-49df-9569-ac383eb42cab-catalog-content\") pod \"certified-operators-m5rft\" (UID: \"18611d84-22b3-49df-9569-ac383eb42cab\") " pod="openshift-marketplace/certified-operators-m5rft" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.460314 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18611d84-22b3-49df-9569-ac383eb42cab-utilities\") pod \"certified-operators-m5rft\" (UID: \"18611d84-22b3-49df-9569-ac383eb42cab\") " pod="openshift-marketplace/certified-operators-m5rft" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.460349 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59kjl\" (UniqueName: \"kubernetes.io/projected/18611d84-22b3-49df-9569-ac383eb42cab-kube-api-access-59kjl\") pod \"certified-operators-m5rft\" (UID: \"18611d84-22b3-49df-9569-ac383eb42cab\") " pod="openshift-marketplace/certified-operators-m5rft" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.461517 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18611d84-22b3-49df-9569-ac383eb42cab-utilities\") pod \"certified-operators-m5rft\" (UID: \"18611d84-22b3-49df-9569-ac383eb42cab\") " pod="openshift-marketplace/certified-operators-m5rft" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.465409 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18611d84-22b3-49df-9569-ac383eb42cab-catalog-content\") pod \"certified-operators-m5rft\" (UID: \"18611d84-22b3-49df-9569-ac383eb42cab\") " pod="openshift-marketplace/certified-operators-m5rft" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.475226 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jw5r7"] Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.476179 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jw5r7" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.482830 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59kjl\" (UniqueName: \"kubernetes.io/projected/18611d84-22b3-49df-9569-ac383eb42cab-kube-api-access-59kjl\") pod \"certified-operators-m5rft\" (UID: \"18611d84-22b3-49df-9569-ac383eb42cab\") " pod="openshift-marketplace/certified-operators-m5rft" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.485906 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jw5r7"] Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.562302 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ced68fc-a805-4595-8043-ce7ec0b802a3-utilities\") pod \"community-operators-jw5r7\" (UID: \"0ced68fc-a805-4595-8043-ce7ec0b802a3\") " pod="openshift-marketplace/community-operators-jw5r7" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.562442 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ced68fc-a805-4595-8043-ce7ec0b802a3-catalog-content\") pod \"community-operators-jw5r7\" (UID: \"0ced68fc-a805-4595-8043-ce7ec0b802a3\") " pod="openshift-marketplace/community-operators-jw5r7" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.562526 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rc2j\" (UniqueName: \"kubernetes.io/projected/0ced68fc-a805-4595-8043-ce7ec0b802a3-kube-api-access-5rc2j\") pod \"community-operators-jw5r7\" (UID: \"0ced68fc-a805-4595-8043-ce7ec0b802a3\") " pod="openshift-marketplace/community-operators-jw5r7" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.614535 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5rft" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.653276 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-96hns"] Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.664592 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ced68fc-a805-4595-8043-ce7ec0b802a3-catalog-content\") pod \"community-operators-jw5r7\" (UID: \"0ced68fc-a805-4595-8043-ce7ec0b802a3\") " pod="openshift-marketplace/community-operators-jw5r7" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.664660 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rc2j\" (UniqueName: \"kubernetes.io/projected/0ced68fc-a805-4595-8043-ce7ec0b802a3-kube-api-access-5rc2j\") pod \"community-operators-jw5r7\" (UID: \"0ced68fc-a805-4595-8043-ce7ec0b802a3\") " pod="openshift-marketplace/community-operators-jw5r7" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.664691 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ced68fc-a805-4595-8043-ce7ec0b802a3-utilities\") pod \"community-operators-jw5r7\" (UID: \"0ced68fc-a805-4595-8043-ce7ec0b802a3\") " pod="openshift-marketplace/community-operators-jw5r7" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.665288 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ced68fc-a805-4595-8043-ce7ec0b802a3-utilities\") pod \"community-operators-jw5r7\" (UID: \"0ced68fc-a805-4595-8043-ce7ec0b802a3\") " pod="openshift-marketplace/community-operators-jw5r7" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.665509 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ced68fc-a805-4595-8043-ce7ec0b802a3-catalog-content\") pod \"community-operators-jw5r7\" (UID: \"0ced68fc-a805-4595-8043-ce7ec0b802a3\") " pod="openshift-marketplace/community-operators-jw5r7" Feb 22 00:09:02 crc kubenswrapper[4857]: W0222 00:09:02.682204 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e0d9776_fb39_4996_9d7d_48a045cd2da9.slice/crio-502c0c4366db02ccc0c067d724cdcc28c9fcc436a802b1579f1cc78658e56c82 WatchSource:0}: Error finding container 502c0c4366db02ccc0c067d724cdcc28c9fcc436a802b1579f1cc78658e56c82: Status 404 returned error can't find the container with id 502c0c4366db02ccc0c067d724cdcc28c9fcc436a802b1579f1cc78658e56c82 Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.700327 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5rpdr"] Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.705817 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rc2j\" (UniqueName: \"kubernetes.io/projected/0ced68fc-a805-4595-8043-ce7ec0b802a3-kube-api-access-5rc2j\") pod \"community-operators-jw5r7\" (UID: \"0ced68fc-a805-4595-8043-ce7ec0b802a3\") " pod="openshift-marketplace/community-operators-jw5r7" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.718260 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nqp5c"] Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.719313 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqp5c" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.731995 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nqp5c"] Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.769595 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d905b043-15bc-4147-8745-d7c26f087559-utilities\") pod \"certified-operators-nqp5c\" (UID: \"d905b043-15bc-4147-8745-d7c26f087559\") " pod="openshift-marketplace/certified-operators-nqp5c" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.769658 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d905b043-15bc-4147-8745-d7c26f087559-catalog-content\") pod \"certified-operators-nqp5c\" (UID: \"d905b043-15bc-4147-8745-d7c26f087559\") " pod="openshift-marketplace/certified-operators-nqp5c" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.769685 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjn4d\" (UniqueName: \"kubernetes.io/projected/d905b043-15bc-4147-8745-d7c26f087559-kube-api-access-pjn4d\") pod \"certified-operators-nqp5c\" (UID: \"d905b043-15bc-4147-8745-d7c26f087559\") " pod="openshift-marketplace/certified-operators-nqp5c" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.792298 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jw5r7" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.872349 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d905b043-15bc-4147-8745-d7c26f087559-utilities\") pod \"certified-operators-nqp5c\" (UID: \"d905b043-15bc-4147-8745-d7c26f087559\") " pod="openshift-marketplace/certified-operators-nqp5c" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.872424 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d905b043-15bc-4147-8745-d7c26f087559-catalog-content\") pod \"certified-operators-nqp5c\" (UID: \"d905b043-15bc-4147-8745-d7c26f087559\") " pod="openshift-marketplace/certified-operators-nqp5c" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.872450 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjn4d\" (UniqueName: \"kubernetes.io/projected/d905b043-15bc-4147-8745-d7c26f087559-kube-api-access-pjn4d\") pod \"certified-operators-nqp5c\" (UID: \"d905b043-15bc-4147-8745-d7c26f087559\") " pod="openshift-marketplace/certified-operators-nqp5c" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.873125 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d905b043-15bc-4147-8745-d7c26f087559-utilities\") pod \"certified-operators-nqp5c\" (UID: \"d905b043-15bc-4147-8745-d7c26f087559\") " pod="openshift-marketplace/certified-operators-nqp5c" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.873337 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d905b043-15bc-4147-8745-d7c26f087559-catalog-content\") pod \"certified-operators-nqp5c\" (UID: \"d905b043-15bc-4147-8745-d7c26f087559\") " pod="openshift-marketplace/certified-operators-nqp5c" Feb 22 00:09:02 crc kubenswrapper[4857]: I0222 00:09:02.910003 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjn4d\" (UniqueName: \"kubernetes.io/projected/d905b043-15bc-4147-8745-d7c26f087559-kube-api-access-pjn4d\") pod \"certified-operators-nqp5c\" (UID: \"d905b043-15bc-4147-8745-d7c26f087559\") " pod="openshift-marketplace/certified-operators-nqp5c" Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.001747 4857 patch_prober.go:28] interesting pod/router-default-5444994796-lfd8v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 22 00:09:03 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Feb 22 00:09:03 crc kubenswrapper[4857]: [+]process-running ok Feb 22 00:09:03 crc kubenswrapper[4857]: healthz check failed Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.001791 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfd8v" podUID="1d0786e3-c21c-46f5-8ca9-7cad8ac62194" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.026727 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m5rft"] Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.053173 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqp5c" Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.116770 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.118999 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jw5r7"] Feb 22 00:09:03 crc kubenswrapper[4857]: W0222 00:09:03.167266 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ced68fc_a805_4595_8043_ce7ec0b802a3.slice/crio-c3bce4f9f27aae2e65ffc5b59da7d061629a0e0a91b9cead4d28f75a73e2e23c WatchSource:0}: Error finding container c3bce4f9f27aae2e65ffc5b59da7d061629a0e0a91b9cead4d28f75a73e2e23c: Status 404 returned error can't find the container with id c3bce4f9f27aae2e65ffc5b59da7d061629a0e0a91b9cead4d28f75a73e2e23c Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.181737 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-96hns" event={"ID":"9e0d9776-fb39-4996-9d7d-48a045cd2da9","Type":"ContainerStarted","Data":"e27f63e562967698d8fca9209ff752f1c5af30fe5cf38c142dde56d26e50bf53"} Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.181779 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-96hns" event={"ID":"9e0d9776-fb39-4996-9d7d-48a045cd2da9","Type":"ContainerStarted","Data":"502c0c4366db02ccc0c067d724cdcc28c9fcc436a802b1579f1cc78658e56c82"} Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.182448 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.189587 4857 generic.go:334] "Generic (PLEG): container finished" podID="05ee3f6d-2b50-4da4-8688-01ae8364d166" containerID="05e9778d99c15d2c5cbae7799ef80b8be54fdda4bbc44e1549eae4965a464f1c" exitCode=0 Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.189667 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" event={"ID":"05ee3f6d-2b50-4da4-8688-01ae8364d166","Type":"ContainerDied","Data":"05e9778d99c15d2c5cbae7799ef80b8be54fdda4bbc44e1549eae4965a464f1c"} Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.191075 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jw5r7" event={"ID":"0ced68fc-a805-4595-8043-ce7ec0b802a3","Type":"ContainerStarted","Data":"c3bce4f9f27aae2e65ffc5b59da7d061629a0e0a91b9cead4d28f75a73e2e23c"} Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.192160 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5rft" event={"ID":"18611d84-22b3-49df-9569-ac383eb42cab","Type":"ContainerStarted","Data":"d16007b43d7e06f1119c6c560c28310c5bed58bc9a79575ffb0b2c94354c1c33"} Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.193551 4857 generic.go:334] "Generic (PLEG): container finished" podID="fa45663f-d64c-431c-be1d-29175cc804ea" containerID="696eb6717c35018debe86fc103767766044bac46fddb0ca57aa1307661843eda" exitCode=0 Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.193614 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rpdr" event={"ID":"fa45663f-d64c-431c-be1d-29175cc804ea","Type":"ContainerDied","Data":"696eb6717c35018debe86fc103767766044bac46fddb0ca57aa1307661843eda"} Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.193634 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rpdr" event={"ID":"fa45663f-d64c-431c-be1d-29175cc804ea","Type":"ContainerStarted","Data":"64daa7086566ca522beeda500051a89d0e4adc6bc1b84155280bd8a2178c446c"} Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.196808 4857 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.205906 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-96hns" podStartSLOduration=126.20588776 podStartE2EDuration="2m6.20588776s" podCreationTimestamp="2026-02-22 00:06:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:09:03.19743788 +0000 UTC m=+150.836167133" watchObservedRunningTime="2026-02-22 00:09:03.20588776 +0000 UTC m=+150.844617013" Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.208106 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-9x24j" event={"ID":"7b02ac9a-c64c-41aa-b2ef-dbff783bf797","Type":"ContainerStarted","Data":"b924c6f09fd78de9206e8d52dc5648bef8590cb05ea0f6ff165af611b48705dd"} Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.308748 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-9x24j" podStartSLOduration=11.308733332 podStartE2EDuration="11.308733332s" podCreationTimestamp="2026-02-22 00:08:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:09:03.307763074 +0000 UTC m=+150.946492337" watchObservedRunningTime="2026-02-22 00:09:03.308733332 +0000 UTC m=+150.947462585" Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.393671 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nqp5c"] Feb 22 00:09:03 crc kubenswrapper[4857]: W0222 00:09:03.480593 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd905b043_15bc_4147_8745_d7c26f087559.slice/crio-0d86d664b271e284c5aaf5f21b227c181bdaf0670ccbee8904682752eb20bc67 WatchSource:0}: Error finding container 0d86d664b271e284c5aaf5f21b227c181bdaf0670ccbee8904682752eb20bc67: Status 404 returned error can't find the container with id 0d86d664b271e284c5aaf5f21b227c181bdaf0670ccbee8904682752eb20bc67 Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.713065 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.713717 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.715217 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.715363 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.719525 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.806339 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/112060a6-9baa-411b-9e8f-9f57c271a1d9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"112060a6-9baa-411b-9e8f-9f57c271a1d9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.806431 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/112060a6-9baa-411b-9e8f-9f57c271a1d9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"112060a6-9baa-411b-9e8f-9f57c271a1d9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.907129 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/112060a6-9baa-411b-9e8f-9f57c271a1d9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"112060a6-9baa-411b-9e8f-9f57c271a1d9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.907215 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/112060a6-9baa-411b-9e8f-9f57c271a1d9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"112060a6-9baa-411b-9e8f-9f57c271a1d9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.907296 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/112060a6-9baa-411b-9e8f-9f57c271a1d9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"112060a6-9baa-411b-9e8f-9f57c271a1d9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 22 00:09:03 crc kubenswrapper[4857]: I0222 00:09:03.929779 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/112060a6-9baa-411b-9e8f-9f57c271a1d9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"112060a6-9baa-411b-9e8f-9f57c271a1d9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.000705 4857 patch_prober.go:28] interesting pod/router-default-5444994796-lfd8v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 22 00:09:04 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Feb 22 00:09:04 crc kubenswrapper[4857]: [+]process-running ok Feb 22 00:09:04 crc kubenswrapper[4857]: healthz check failed Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.000791 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfd8v" podUID="1d0786e3-c21c-46f5-8ca9-7cad8ac62194" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.032369 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.210086 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.213818 4857 generic.go:334] "Generic (PLEG): container finished" podID="0ced68fc-a805-4595-8043-ce7ec0b802a3" containerID="13ccc6e155863d7c55fdbe57ee7e38a666c9aba35c96187baa7d457023787ffa" exitCode=0 Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.213886 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jw5r7" event={"ID":"0ced68fc-a805-4595-8043-ce7ec0b802a3","Type":"ContainerDied","Data":"13ccc6e155863d7c55fdbe57ee7e38a666c9aba35c96187baa7d457023787ffa"} Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.216310 4857 generic.go:334] "Generic (PLEG): container finished" podID="18611d84-22b3-49df-9569-ac383eb42cab" containerID="ee5d8876d1488183eb5edf35a9efa3293b59bfcb8d52c5fa2743306c94e399f5" exitCode=0 Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.216376 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5rft" event={"ID":"18611d84-22b3-49df-9569-ac383eb42cab","Type":"ContainerDied","Data":"ee5d8876d1488183eb5edf35a9efa3293b59bfcb8d52c5fa2743306c94e399f5"} Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.228606 4857 generic.go:334] "Generic (PLEG): container finished" podID="d905b043-15bc-4147-8745-d7c26f087559" containerID="2ca904adbfa7bfd8a4cc01573aaf71122c1db1d1c068aa74d09bcebac7377ead" exitCode=0 Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.229159 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqp5c" event={"ID":"d905b043-15bc-4147-8745-d7c26f087559","Type":"ContainerDied","Data":"2ca904adbfa7bfd8a4cc01573aaf71122c1db1d1c068aa74d09bcebac7377ead"} Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.229211 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqp5c" event={"ID":"d905b043-15bc-4147-8745-d7c26f087559","Type":"ContainerStarted","Data":"0d86d664b271e284c5aaf5f21b227c181bdaf0670ccbee8904682752eb20bc67"} Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.283100 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-797dx"] Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.290338 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-797dx" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.296194 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.302566 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-797dx"] Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.415804 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f2d337-12f1-4949-9872-08df9aa084ab-utilities\") pod \"redhat-marketplace-797dx\" (UID: \"10f2d337-12f1-4949-9872-08df9aa084ab\") " pod="openshift-marketplace/redhat-marketplace-797dx" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.415872 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f2d337-12f1-4949-9872-08df9aa084ab-catalog-content\") pod \"redhat-marketplace-797dx\" (UID: \"10f2d337-12f1-4949-9872-08df9aa084ab\") " pod="openshift-marketplace/redhat-marketplace-797dx" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.415914 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8v27\" (UniqueName: \"kubernetes.io/projected/10f2d337-12f1-4949-9872-08df9aa084ab-kube-api-access-f8v27\") pod \"redhat-marketplace-797dx\" (UID: \"10f2d337-12f1-4949-9872-08df9aa084ab\") " pod="openshift-marketplace/redhat-marketplace-797dx" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.446241 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.516319 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05ee3f6d-2b50-4da4-8688-01ae8364d166-secret-volume\") pod \"05ee3f6d-2b50-4da4-8688-01ae8364d166\" (UID: \"05ee3f6d-2b50-4da4-8688-01ae8364d166\") " Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.516449 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57hqs\" (UniqueName: \"kubernetes.io/projected/05ee3f6d-2b50-4da4-8688-01ae8364d166-kube-api-access-57hqs\") pod \"05ee3f6d-2b50-4da4-8688-01ae8364d166\" (UID: \"05ee3f6d-2b50-4da4-8688-01ae8364d166\") " Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.516485 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05ee3f6d-2b50-4da4-8688-01ae8364d166-config-volume\") pod \"05ee3f6d-2b50-4da4-8688-01ae8364d166\" (UID: \"05ee3f6d-2b50-4da4-8688-01ae8364d166\") " Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.516651 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f2d337-12f1-4949-9872-08df9aa084ab-utilities\") pod \"redhat-marketplace-797dx\" (UID: \"10f2d337-12f1-4949-9872-08df9aa084ab\") " pod="openshift-marketplace/redhat-marketplace-797dx" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.516691 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f2d337-12f1-4949-9872-08df9aa084ab-catalog-content\") pod \"redhat-marketplace-797dx\" (UID: \"10f2d337-12f1-4949-9872-08df9aa084ab\") " pod="openshift-marketplace/redhat-marketplace-797dx" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.516738 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8v27\" (UniqueName: \"kubernetes.io/projected/10f2d337-12f1-4949-9872-08df9aa084ab-kube-api-access-f8v27\") pod \"redhat-marketplace-797dx\" (UID: \"10f2d337-12f1-4949-9872-08df9aa084ab\") " pod="openshift-marketplace/redhat-marketplace-797dx" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.517521 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05ee3f6d-2b50-4da4-8688-01ae8364d166-config-volume" (OuterVolumeSpecName: "config-volume") pod "05ee3f6d-2b50-4da4-8688-01ae8364d166" (UID: "05ee3f6d-2b50-4da4-8688-01ae8364d166"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.517995 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f2d337-12f1-4949-9872-08df9aa084ab-utilities\") pod \"redhat-marketplace-797dx\" (UID: \"10f2d337-12f1-4949-9872-08df9aa084ab\") " pod="openshift-marketplace/redhat-marketplace-797dx" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.518021 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f2d337-12f1-4949-9872-08df9aa084ab-catalog-content\") pod \"redhat-marketplace-797dx\" (UID: \"10f2d337-12f1-4949-9872-08df9aa084ab\") " pod="openshift-marketplace/redhat-marketplace-797dx" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.521502 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05ee3f6d-2b50-4da4-8688-01ae8364d166-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "05ee3f6d-2b50-4da4-8688-01ae8364d166" (UID: "05ee3f6d-2b50-4da4-8688-01ae8364d166"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.521937 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05ee3f6d-2b50-4da4-8688-01ae8364d166-kube-api-access-57hqs" (OuterVolumeSpecName: "kube-api-access-57hqs") pod "05ee3f6d-2b50-4da4-8688-01ae8364d166" (UID: "05ee3f6d-2b50-4da4-8688-01ae8364d166"). InnerVolumeSpecName "kube-api-access-57hqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.531052 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8v27\" (UniqueName: \"kubernetes.io/projected/10f2d337-12f1-4949-9872-08df9aa084ab-kube-api-access-f8v27\") pod \"redhat-marketplace-797dx\" (UID: \"10f2d337-12f1-4949-9872-08df9aa084ab\") " pod="openshift-marketplace/redhat-marketplace-797dx" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.579528 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.579588 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.581195 4857 patch_prober.go:28] interesting pod/console-f9d7485db-wq8b8 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.581258 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-wq8b8" podUID="f5acab23-9994-4c18-bc50-4f5ec7c7f525" containerName="console" probeResult="failure" output="Get \"https://10.217.0.23:8443/health\": dial tcp 10.217.0.23:8443: connect: connection refused" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.617214 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-797dx" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.619061 4857 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05ee3f6d-2b50-4da4-8688-01ae8364d166-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.619093 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57hqs\" (UniqueName: \"kubernetes.io/projected/05ee3f6d-2b50-4da4-8688-01ae8364d166-kube-api-access-57hqs\") on node \"crc\" DevicePath \"\"" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.619103 4857 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05ee3f6d-2b50-4da4-8688-01ae8364d166-config-volume\") on node \"crc\" DevicePath \"\"" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.679814 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bdmrq"] Feb 22 00:09:04 crc kubenswrapper[4857]: E0222 00:09:04.680104 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05ee3f6d-2b50-4da4-8688-01ae8364d166" containerName="collect-profiles" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.680119 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="05ee3f6d-2b50-4da4-8688-01ae8364d166" containerName="collect-profiles" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.680259 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="05ee3f6d-2b50-4da4-8688-01ae8364d166" containerName="collect-profiles" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.681281 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bdmrq" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.696005 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bdmrq"] Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.719647 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c14592b-a2d9-46c6-b818-6da2942cb895-catalog-content\") pod \"redhat-marketplace-bdmrq\" (UID: \"0c14592b-a2d9-46c6-b818-6da2942cb895\") " pod="openshift-marketplace/redhat-marketplace-bdmrq" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.719881 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c14592b-a2d9-46c6-b818-6da2942cb895-utilities\") pod \"redhat-marketplace-bdmrq\" (UID: \"0c14592b-a2d9-46c6-b818-6da2942cb895\") " pod="openshift-marketplace/redhat-marketplace-bdmrq" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.719946 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ml52g\" (UniqueName: \"kubernetes.io/projected/0c14592b-a2d9-46c6-b818-6da2942cb895-kube-api-access-ml52g\") pod \"redhat-marketplace-bdmrq\" (UID: \"0c14592b-a2d9-46c6-b818-6da2942cb895\") " pod="openshift-marketplace/redhat-marketplace-bdmrq" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.772797 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.773853 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.778840 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.821416 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c14592b-a2d9-46c6-b818-6da2942cb895-catalog-content\") pod \"redhat-marketplace-bdmrq\" (UID: \"0c14592b-a2d9-46c6-b818-6da2942cb895\") " pod="openshift-marketplace/redhat-marketplace-bdmrq" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.821508 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c14592b-a2d9-46c6-b818-6da2942cb895-utilities\") pod \"redhat-marketplace-bdmrq\" (UID: \"0c14592b-a2d9-46c6-b818-6da2942cb895\") " pod="openshift-marketplace/redhat-marketplace-bdmrq" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.821733 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml52g\" (UniqueName: \"kubernetes.io/projected/0c14592b-a2d9-46c6-b818-6da2942cb895-kube-api-access-ml52g\") pod \"redhat-marketplace-bdmrq\" (UID: \"0c14592b-a2d9-46c6-b818-6da2942cb895\") " pod="openshift-marketplace/redhat-marketplace-bdmrq" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.821973 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c14592b-a2d9-46c6-b818-6da2942cb895-catalog-content\") pod \"redhat-marketplace-bdmrq\" (UID: \"0c14592b-a2d9-46c6-b818-6da2942cb895\") " pod="openshift-marketplace/redhat-marketplace-bdmrq" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.821505 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-797dx"] Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.823122 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c14592b-a2d9-46c6-b818-6da2942cb895-utilities\") pod \"redhat-marketplace-bdmrq\" (UID: \"0c14592b-a2d9-46c6-b818-6da2942cb895\") " pod="openshift-marketplace/redhat-marketplace-bdmrq" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.841121 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml52g\" (UniqueName: \"kubernetes.io/projected/0c14592b-a2d9-46c6-b818-6da2942cb895-kube-api-access-ml52g\") pod \"redhat-marketplace-bdmrq\" (UID: \"0c14592b-a2d9-46c6-b818-6da2942cb895\") " pod="openshift-marketplace/redhat-marketplace-bdmrq" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.995844 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bdmrq" Feb 22 00:09:04 crc kubenswrapper[4857]: I0222 00:09:04.997649 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.001534 4857 patch_prober.go:28] interesting pod/router-default-5444994796-lfd8v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 22 00:09:05 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Feb 22 00:09:05 crc kubenswrapper[4857]: [+]process-running ok Feb 22 00:09:05 crc kubenswrapper[4857]: healthz check failed Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.001586 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfd8v" podUID="1d0786e3-c21c-46f5-8ca9-7cad8ac62194" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.025477 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nqshz" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.034593 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.035411 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.037570 4857 patch_prober.go:28] interesting pod/downloads-7954f5f757-h8pq2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.037581 4857 patch_prober.go:28] interesting pod/downloads-7954f5f757-h8pq2 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.037616 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-h8pq2" podUID="ff701efb-cabc-479a-ac05-4e9334e70385" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.037627 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-h8pq2" podUID="ff701efb-cabc-479a-ac05-4e9334e70385" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.038657 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.038895 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.052557 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.125468 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/feceeec5-56fd-4977-9f97-f0dc03546b98-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"feceeec5-56fd-4977-9f97-f0dc03546b98\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.125839 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/feceeec5-56fd-4977-9f97-f0dc03546b98-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"feceeec5-56fd-4977-9f97-f0dc03546b98\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.208608 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bdmrq"] Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.226949 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/feceeec5-56fd-4977-9f97-f0dc03546b98-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"feceeec5-56fd-4977-9f97-f0dc03546b98\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.227006 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/feceeec5-56fd-4977-9f97-f0dc03546b98-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"feceeec5-56fd-4977-9f97-f0dc03546b98\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.227095 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/feceeec5-56fd-4977-9f97-f0dc03546b98-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"feceeec5-56fd-4977-9f97-f0dc03546b98\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.234281 4857 generic.go:334] "Generic (PLEG): container finished" podID="10f2d337-12f1-4949-9872-08df9aa084ab" containerID="c035cf724c76965a182f07533df88b7f93333062ee13f7d63bd2843ba86a7173" exitCode=0 Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.234511 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-797dx" event={"ID":"10f2d337-12f1-4949-9872-08df9aa084ab","Type":"ContainerDied","Data":"c035cf724c76965a182f07533df88b7f93333062ee13f7d63bd2843ba86a7173"} Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.234542 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-797dx" event={"ID":"10f2d337-12f1-4949-9872-08df9aa084ab","Type":"ContainerStarted","Data":"702273ed60bd457682a58cb48d335b676624d61f9b8253808e3653e3837389ea"} Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.236176 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.236187 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29528640-sbzs9" event={"ID":"05ee3f6d-2b50-4da4-8688-01ae8364d166","Type":"ContainerDied","Data":"13a06d02575b7ac3689f040690d6ca2ca539e333164983aab2bb47c4982f5268"} Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.236227 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13a06d02575b7ac3689f040690d6ca2ca539e333164983aab2bb47c4982f5268" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.239315 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bdmrq" event={"ID":"0c14592b-a2d9-46c6-b818-6da2942cb895","Type":"ContainerStarted","Data":"9ea747affc4bee529a5ed87989e12c5fd6a9b2f6c28e90b35077f68c3825e094"} Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.250131 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/feceeec5-56fd-4977-9f97-f0dc03546b98-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"feceeec5-56fd-4977-9f97-f0dc03546b98\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.251404 4857 generic.go:334] "Generic (PLEG): container finished" podID="112060a6-9baa-411b-9e8f-9f57c271a1d9" containerID="99a199b1515a4ba5da78a947e1074cf64dbbad869e505f78a13dcc999fc64a3c" exitCode=0 Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.251486 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"112060a6-9baa-411b-9e8f-9f57c271a1d9","Type":"ContainerDied","Data":"99a199b1515a4ba5da78a947e1074cf64dbbad869e505f78a13dcc999fc64a3c"} Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.251516 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"112060a6-9baa-411b-9e8f-9f57c271a1d9","Type":"ContainerStarted","Data":"92f34d1000614d8aa5eb190971d740e51451453b6b4d387e6f598e31196f7727"} Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.258339 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-mx94d" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.313537 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xd4g7"] Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.319820 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xd4g7"] Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.320315 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xd4g7" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.324762 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.364287 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.431175 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd2qp\" (UniqueName: \"kubernetes.io/projected/6d73807b-5e0c-4607-8393-5123a6cdbf9b-kube-api-access-cd2qp\") pod \"redhat-operators-xd4g7\" (UID: \"6d73807b-5e0c-4607-8393-5123a6cdbf9b\") " pod="openshift-marketplace/redhat-operators-xd4g7" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.431351 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d73807b-5e0c-4607-8393-5123a6cdbf9b-utilities\") pod \"redhat-operators-xd4g7\" (UID: \"6d73807b-5e0c-4607-8393-5123a6cdbf9b\") " pod="openshift-marketplace/redhat-operators-xd4g7" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.431519 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d73807b-5e0c-4607-8393-5123a6cdbf9b-catalog-content\") pod \"redhat-operators-xd4g7\" (UID: \"6d73807b-5e0c-4607-8393-5123a6cdbf9b\") " pod="openshift-marketplace/redhat-operators-xd4g7" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.532693 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd2qp\" (UniqueName: \"kubernetes.io/projected/6d73807b-5e0c-4607-8393-5123a6cdbf9b-kube-api-access-cd2qp\") pod \"redhat-operators-xd4g7\" (UID: \"6d73807b-5e0c-4607-8393-5123a6cdbf9b\") " pod="openshift-marketplace/redhat-operators-xd4g7" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.532910 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d73807b-5e0c-4607-8393-5123a6cdbf9b-utilities\") pod \"redhat-operators-xd4g7\" (UID: \"6d73807b-5e0c-4607-8393-5123a6cdbf9b\") " pod="openshift-marketplace/redhat-operators-xd4g7" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.533093 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d73807b-5e0c-4607-8393-5123a6cdbf9b-catalog-content\") pod \"redhat-operators-xd4g7\" (UID: \"6d73807b-5e0c-4607-8393-5123a6cdbf9b\") " pod="openshift-marketplace/redhat-operators-xd4g7" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.533930 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d73807b-5e0c-4607-8393-5123a6cdbf9b-utilities\") pod \"redhat-operators-xd4g7\" (UID: \"6d73807b-5e0c-4607-8393-5123a6cdbf9b\") " pod="openshift-marketplace/redhat-operators-xd4g7" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.534001 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d73807b-5e0c-4607-8393-5123a6cdbf9b-catalog-content\") pod \"redhat-operators-xd4g7\" (UID: \"6d73807b-5e0c-4607-8393-5123a6cdbf9b\") " pod="openshift-marketplace/redhat-operators-xd4g7" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.566179 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.572875 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd2qp\" (UniqueName: \"kubernetes.io/projected/6d73807b-5e0c-4607-8393-5123a6cdbf9b-kube-api-access-cd2qp\") pod \"redhat-operators-xd4g7\" (UID: \"6d73807b-5e0c-4607-8393-5123a6cdbf9b\") " pod="openshift-marketplace/redhat-operators-xd4g7" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.648366 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xd4g7" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.669813 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.676344 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fmz46"] Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.687586 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fmz46" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.693284 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fmz46"] Feb 22 00:09:05 crc kubenswrapper[4857]: W0222 00:09:05.702448 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podfeceeec5_56fd_4977_9f97_f0dc03546b98.slice/crio-bfb7a47fae5a78f07e4628feb4c5727306e6c1cf3fa0cbd0ecc2a7d9808e5792 WatchSource:0}: Error finding container bfb7a47fae5a78f07e4628feb4c5727306e6c1cf3fa0cbd0ecc2a7d9808e5792: Status 404 returned error can't find the container with id bfb7a47fae5a78f07e4628feb4c5727306e6c1cf3fa0cbd0ecc2a7d9808e5792 Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.735677 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/299c4a35-6470-4a40-9b90-ad0390497677-utilities\") pod \"redhat-operators-fmz46\" (UID: \"299c4a35-6470-4a40-9b90-ad0390497677\") " pod="openshift-marketplace/redhat-operators-fmz46" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.736006 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9fqw\" (UniqueName: \"kubernetes.io/projected/299c4a35-6470-4a40-9b90-ad0390497677-kube-api-access-k9fqw\") pod \"redhat-operators-fmz46\" (UID: \"299c4a35-6470-4a40-9b90-ad0390497677\") " pod="openshift-marketplace/redhat-operators-fmz46" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.736049 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/299c4a35-6470-4a40-9b90-ad0390497677-catalog-content\") pod \"redhat-operators-fmz46\" (UID: \"299c4a35-6470-4a40-9b90-ad0390497677\") " pod="openshift-marketplace/redhat-operators-fmz46" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.837417 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/299c4a35-6470-4a40-9b90-ad0390497677-utilities\") pod \"redhat-operators-fmz46\" (UID: \"299c4a35-6470-4a40-9b90-ad0390497677\") " pod="openshift-marketplace/redhat-operators-fmz46" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.837881 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9fqw\" (UniqueName: \"kubernetes.io/projected/299c4a35-6470-4a40-9b90-ad0390497677-kube-api-access-k9fqw\") pod \"redhat-operators-fmz46\" (UID: \"299c4a35-6470-4a40-9b90-ad0390497677\") " pod="openshift-marketplace/redhat-operators-fmz46" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.837909 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/299c4a35-6470-4a40-9b90-ad0390497677-catalog-content\") pod \"redhat-operators-fmz46\" (UID: \"299c4a35-6470-4a40-9b90-ad0390497677\") " pod="openshift-marketplace/redhat-operators-fmz46" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.838159 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/299c4a35-6470-4a40-9b90-ad0390497677-catalog-content\") pod \"redhat-operators-fmz46\" (UID: \"299c4a35-6470-4a40-9b90-ad0390497677\") " pod="openshift-marketplace/redhat-operators-fmz46" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.837816 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/299c4a35-6470-4a40-9b90-ad0390497677-utilities\") pod \"redhat-operators-fmz46\" (UID: \"299c4a35-6470-4a40-9b90-ad0390497677\") " pod="openshift-marketplace/redhat-operators-fmz46" Feb 22 00:09:05 crc kubenswrapper[4857]: I0222 00:09:05.869398 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9fqw\" (UniqueName: \"kubernetes.io/projected/299c4a35-6470-4a40-9b90-ad0390497677-kube-api-access-k9fqw\") pod \"redhat-operators-fmz46\" (UID: \"299c4a35-6470-4a40-9b90-ad0390497677\") " pod="openshift-marketplace/redhat-operators-fmz46" Feb 22 00:09:06 crc kubenswrapper[4857]: I0222 00:09:06.002394 4857 patch_prober.go:28] interesting pod/router-default-5444994796-lfd8v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 22 00:09:06 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Feb 22 00:09:06 crc kubenswrapper[4857]: [+]process-running ok Feb 22 00:09:06 crc kubenswrapper[4857]: healthz check failed Feb 22 00:09:06 crc kubenswrapper[4857]: I0222 00:09:06.002496 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfd8v" podUID="1d0786e3-c21c-46f5-8ca9-7cad8ac62194" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 22 00:09:06 crc kubenswrapper[4857]: I0222 00:09:06.021650 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fmz46" Feb 22 00:09:06 crc kubenswrapper[4857]: I0222 00:09:06.097165 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xd4g7"] Feb 22 00:09:06 crc kubenswrapper[4857]: W0222 00:09:06.115000 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d73807b_5e0c_4607_8393_5123a6cdbf9b.slice/crio-f53f3c846ec1ce2552bb9b740f4929b3c4292716d057a1378b9f20362d315934 WatchSource:0}: Error finding container f53f3c846ec1ce2552bb9b740f4929b3c4292716d057a1378b9f20362d315934: Status 404 returned error can't find the container with id f53f3c846ec1ce2552bb9b740f4929b3c4292716d057a1378b9f20362d315934 Feb 22 00:09:06 crc kubenswrapper[4857]: I0222 00:09:06.292094 4857 generic.go:334] "Generic (PLEG): container finished" podID="0c14592b-a2d9-46c6-b818-6da2942cb895" containerID="aa9259c83a2418aa714903102d753ff124b035ff728274241cc33a61175681a1" exitCode=0 Feb 22 00:09:06 crc kubenswrapper[4857]: I0222 00:09:06.292305 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bdmrq" event={"ID":"0c14592b-a2d9-46c6-b818-6da2942cb895","Type":"ContainerDied","Data":"aa9259c83a2418aa714903102d753ff124b035ff728274241cc33a61175681a1"} Feb 22 00:09:06 crc kubenswrapper[4857]: I0222 00:09:06.297466 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"feceeec5-56fd-4977-9f97-f0dc03546b98","Type":"ContainerStarted","Data":"eb9ce56053ad8b72bc49fbd4434d23a591e626fd676f84209b0065e3399061a5"} Feb 22 00:09:06 crc kubenswrapper[4857]: I0222 00:09:06.297503 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"feceeec5-56fd-4977-9f97-f0dc03546b98","Type":"ContainerStarted","Data":"bfb7a47fae5a78f07e4628feb4c5727306e6c1cf3fa0cbd0ecc2a7d9808e5792"} Feb 22 00:09:06 crc kubenswrapper[4857]: I0222 00:09:06.300139 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xd4g7" event={"ID":"6d73807b-5e0c-4607-8393-5123a6cdbf9b","Type":"ContainerStarted","Data":"f53f3c846ec1ce2552bb9b740f4929b3c4292716d057a1378b9f20362d315934"} Feb 22 00:09:06 crc kubenswrapper[4857]: I0222 00:09:06.331318 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=1.331300767 podStartE2EDuration="1.331300767s" podCreationTimestamp="2026-02-22 00:09:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:09:06.330813383 +0000 UTC m=+153.969542636" watchObservedRunningTime="2026-02-22 00:09:06.331300767 +0000 UTC m=+153.970030020" Feb 22 00:09:06 crc kubenswrapper[4857]: I0222 00:09:06.591860 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fmz46"] Feb 22 00:09:06 crc kubenswrapper[4857]: I0222 00:09:06.599668 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 22 00:09:06 crc kubenswrapper[4857]: W0222 00:09:06.620329 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod299c4a35_6470_4a40_9b90_ad0390497677.slice/crio-1f4d049e289f91deafbff62b2081acf0666bd799d7f02c2633ea706901633011 WatchSource:0}: Error finding container 1f4d049e289f91deafbff62b2081acf0666bd799d7f02c2633ea706901633011: Status 404 returned error can't find the container with id 1f4d049e289f91deafbff62b2081acf0666bd799d7f02c2633ea706901633011 Feb 22 00:09:06 crc kubenswrapper[4857]: I0222 00:09:06.652992 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/112060a6-9baa-411b-9e8f-9f57c271a1d9-kube-api-access\") pod \"112060a6-9baa-411b-9e8f-9f57c271a1d9\" (UID: \"112060a6-9baa-411b-9e8f-9f57c271a1d9\") " Feb 22 00:09:06 crc kubenswrapper[4857]: I0222 00:09:06.653392 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/112060a6-9baa-411b-9e8f-9f57c271a1d9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "112060a6-9baa-411b-9e8f-9f57c271a1d9" (UID: "112060a6-9baa-411b-9e8f-9f57c271a1d9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:09:06 crc kubenswrapper[4857]: I0222 00:09:06.653633 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/112060a6-9baa-411b-9e8f-9f57c271a1d9-kubelet-dir\") pod \"112060a6-9baa-411b-9e8f-9f57c271a1d9\" (UID: \"112060a6-9baa-411b-9e8f-9f57c271a1d9\") " Feb 22 00:09:06 crc kubenswrapper[4857]: I0222 00:09:06.654094 4857 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/112060a6-9baa-411b-9e8f-9f57c271a1d9-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 22 00:09:06 crc kubenswrapper[4857]: I0222 00:09:06.659746 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/112060a6-9baa-411b-9e8f-9f57c271a1d9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "112060a6-9baa-411b-9e8f-9f57c271a1d9" (UID: "112060a6-9baa-411b-9e8f-9f57c271a1d9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:09:06 crc kubenswrapper[4857]: I0222 00:09:06.755855 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/112060a6-9baa-411b-9e8f-9f57c271a1d9-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 22 00:09:07 crc kubenswrapper[4857]: I0222 00:09:07.001604 4857 patch_prober.go:28] interesting pod/router-default-5444994796-lfd8v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 22 00:09:07 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Feb 22 00:09:07 crc kubenswrapper[4857]: [+]process-running ok Feb 22 00:09:07 crc kubenswrapper[4857]: healthz check failed Feb 22 00:09:07 crc kubenswrapper[4857]: I0222 00:09:07.001661 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfd8v" podUID="1d0786e3-c21c-46f5-8ca9-7cad8ac62194" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 22 00:09:07 crc kubenswrapper[4857]: I0222 00:09:07.334132 4857 generic.go:334] "Generic (PLEG): container finished" podID="feceeec5-56fd-4977-9f97-f0dc03546b98" containerID="eb9ce56053ad8b72bc49fbd4434d23a591e626fd676f84209b0065e3399061a5" exitCode=0 Feb 22 00:09:07 crc kubenswrapper[4857]: I0222 00:09:07.334968 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"feceeec5-56fd-4977-9f97-f0dc03546b98","Type":"ContainerDied","Data":"eb9ce56053ad8b72bc49fbd4434d23a591e626fd676f84209b0065e3399061a5"} Feb 22 00:09:07 crc kubenswrapper[4857]: I0222 00:09:07.354484 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"112060a6-9baa-411b-9e8f-9f57c271a1d9","Type":"ContainerDied","Data":"92f34d1000614d8aa5eb190971d740e51451453b6b4d387e6f598e31196f7727"} Feb 22 00:09:07 crc kubenswrapper[4857]: I0222 00:09:07.354542 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92f34d1000614d8aa5eb190971d740e51451453b6b4d387e6f598e31196f7727" Feb 22 00:09:07 crc kubenswrapper[4857]: I0222 00:09:07.354611 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 22 00:09:07 crc kubenswrapper[4857]: I0222 00:09:07.373795 4857 generic.go:334] "Generic (PLEG): container finished" podID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" containerID="4ab98083305747c64255ef608e22c65a22fd727e68fb36f8611ab725e7b14ba4" exitCode=0 Feb 22 00:09:07 crc kubenswrapper[4857]: I0222 00:09:07.374367 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xd4g7" event={"ID":"6d73807b-5e0c-4607-8393-5123a6cdbf9b","Type":"ContainerDied","Data":"4ab98083305747c64255ef608e22c65a22fd727e68fb36f8611ab725e7b14ba4"} Feb 22 00:09:07 crc kubenswrapper[4857]: I0222 00:09:07.385632 4857 generic.go:334] "Generic (PLEG): container finished" podID="299c4a35-6470-4a40-9b90-ad0390497677" containerID="5fe027cb577b65b59ae8a4d1e15b4d33b7276edd73f93df2e2a9de478d1cf8cb" exitCode=0 Feb 22 00:09:07 crc kubenswrapper[4857]: I0222 00:09:07.385941 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmz46" event={"ID":"299c4a35-6470-4a40-9b90-ad0390497677","Type":"ContainerDied","Data":"5fe027cb577b65b59ae8a4d1e15b4d33b7276edd73f93df2e2a9de478d1cf8cb"} Feb 22 00:09:07 crc kubenswrapper[4857]: I0222 00:09:07.386167 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmz46" event={"ID":"299c4a35-6470-4a40-9b90-ad0390497677","Type":"ContainerStarted","Data":"1f4d049e289f91deafbff62b2081acf0666bd799d7f02c2633ea706901633011"} Feb 22 00:09:08 crc kubenswrapper[4857]: I0222 00:09:08.000463 4857 patch_prober.go:28] interesting pod/router-default-5444994796-lfd8v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 22 00:09:08 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Feb 22 00:09:08 crc kubenswrapper[4857]: [+]process-running ok Feb 22 00:09:08 crc kubenswrapper[4857]: healthz check failed Feb 22 00:09:08 crc kubenswrapper[4857]: I0222 00:09:08.000522 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfd8v" podUID="1d0786e3-c21c-46f5-8ca9-7cad8ac62194" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 22 00:09:08 crc kubenswrapper[4857]: I0222 00:09:08.796223 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 22 00:09:08 crc kubenswrapper[4857]: I0222 00:09:08.942665 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/feceeec5-56fd-4977-9f97-f0dc03546b98-kube-api-access\") pod \"feceeec5-56fd-4977-9f97-f0dc03546b98\" (UID: \"feceeec5-56fd-4977-9f97-f0dc03546b98\") " Feb 22 00:09:08 crc kubenswrapper[4857]: I0222 00:09:08.942750 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/feceeec5-56fd-4977-9f97-f0dc03546b98-kubelet-dir\") pod \"feceeec5-56fd-4977-9f97-f0dc03546b98\" (UID: \"feceeec5-56fd-4977-9f97-f0dc03546b98\") " Feb 22 00:09:08 crc kubenswrapper[4857]: I0222 00:09:08.943210 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/feceeec5-56fd-4977-9f97-f0dc03546b98-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "feceeec5-56fd-4977-9f97-f0dc03546b98" (UID: "feceeec5-56fd-4977-9f97-f0dc03546b98"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:09:08 crc kubenswrapper[4857]: I0222 00:09:08.965400 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/feceeec5-56fd-4977-9f97-f0dc03546b98-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "feceeec5-56fd-4977-9f97-f0dc03546b98" (UID: "feceeec5-56fd-4977-9f97-f0dc03546b98"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:09:09 crc kubenswrapper[4857]: I0222 00:09:09.001374 4857 patch_prober.go:28] interesting pod/router-default-5444994796-lfd8v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 22 00:09:09 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Feb 22 00:09:09 crc kubenswrapper[4857]: [+]process-running ok Feb 22 00:09:09 crc kubenswrapper[4857]: healthz check failed Feb 22 00:09:09 crc kubenswrapper[4857]: I0222 00:09:09.001431 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfd8v" podUID="1d0786e3-c21c-46f5-8ca9-7cad8ac62194" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 22 00:09:09 crc kubenswrapper[4857]: I0222 00:09:09.044458 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/feceeec5-56fd-4977-9f97-f0dc03546b98-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 22 00:09:09 crc kubenswrapper[4857]: I0222 00:09:09.044487 4857 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/feceeec5-56fd-4977-9f97-f0dc03546b98-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 22 00:09:09 crc kubenswrapper[4857]: I0222 00:09:09.426596 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"feceeec5-56fd-4977-9f97-f0dc03546b98","Type":"ContainerDied","Data":"bfb7a47fae5a78f07e4628feb4c5727306e6c1cf3fa0cbd0ecc2a7d9808e5792"} Feb 22 00:09:09 crc kubenswrapper[4857]: I0222 00:09:09.426643 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfb7a47fae5a78f07e4628feb4c5727306e6c1cf3fa0cbd0ecc2a7d9808e5792" Feb 22 00:09:09 crc kubenswrapper[4857]: I0222 00:09:09.426708 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 22 00:09:10 crc kubenswrapper[4857]: I0222 00:09:10.000914 4857 patch_prober.go:28] interesting pod/router-default-5444994796-lfd8v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 22 00:09:10 crc kubenswrapper[4857]: [-]has-synced failed: reason withheld Feb 22 00:09:10 crc kubenswrapper[4857]: [+]process-running ok Feb 22 00:09:10 crc kubenswrapper[4857]: healthz check failed Feb 22 00:09:10 crc kubenswrapper[4857]: I0222 00:09:10.000958 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfd8v" podUID="1d0786e3-c21c-46f5-8ca9-7cad8ac62194" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 22 00:09:10 crc kubenswrapper[4857]: I0222 00:09:10.432363 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:09:10 crc kubenswrapper[4857]: I0222 00:09:10.432670 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:09:10 crc kubenswrapper[4857]: I0222 00:09:10.647978 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-c4xt9" Feb 22 00:09:11 crc kubenswrapper[4857]: I0222 00:09:11.008067 4857 patch_prober.go:28] interesting pod/router-default-5444994796-lfd8v container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 22 00:09:11 crc kubenswrapper[4857]: [+]has-synced ok Feb 22 00:09:11 crc kubenswrapper[4857]: [+]process-running ok Feb 22 00:09:11 crc kubenswrapper[4857]: healthz check failed Feb 22 00:09:11 crc kubenswrapper[4857]: I0222 00:09:11.008139 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-lfd8v" podUID="1d0786e3-c21c-46f5-8ca9-7cad8ac62194" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 22 00:09:12 crc kubenswrapper[4857]: I0222 00:09:12.004943 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:09:12 crc kubenswrapper[4857]: I0222 00:09:12.007510 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-lfd8v" Feb 22 00:09:14 crc kubenswrapper[4857]: I0222 00:09:14.584347 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:09:14 crc kubenswrapper[4857]: I0222 00:09:14.588229 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-wq8b8" Feb 22 00:09:15 crc kubenswrapper[4857]: I0222 00:09:15.037794 4857 patch_prober.go:28] interesting pod/downloads-7954f5f757-h8pq2 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Feb 22 00:09:15 crc kubenswrapper[4857]: I0222 00:09:15.037843 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-h8pq2" podUID="ff701efb-cabc-479a-ac05-4e9334e70385" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Feb 22 00:09:15 crc kubenswrapper[4857]: I0222 00:09:15.037794 4857 patch_prober.go:28] interesting pod/downloads-7954f5f757-h8pq2 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Feb 22 00:09:15 crc kubenswrapper[4857]: I0222 00:09:15.038158 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-h8pq2" podUID="ff701efb-cabc-479a-ac05-4e9334e70385" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Feb 22 00:09:16 crc kubenswrapper[4857]: I0222 00:09:16.447961 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:09:19 crc kubenswrapper[4857]: I0222 00:09:19.501458 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs\") pod \"network-metrics-daemon-jnm7q\" (UID: \"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\") " pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:09:19 crc kubenswrapper[4857]: I0222 00:09:19.507629 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a0fb0fe8-ad7a-48a5-8a76-737cd5076f53-metrics-certs\") pod \"network-metrics-daemon-jnm7q\" (UID: \"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53\") " pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:09:19 crc kubenswrapper[4857]: I0222 00:09:19.799630 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jnm7q" Feb 22 00:09:22 crc kubenswrapper[4857]: I0222 00:09:22.398329 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:09:25 crc kubenswrapper[4857]: I0222 00:09:25.045259 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-h8pq2" Feb 22 00:09:30 crc kubenswrapper[4857]: I0222 00:09:30.585485 4857 generic.go:334] "Generic (PLEG): container finished" podID="f07c5f6c-743d-4032-b70b-ecc0f514295d" containerID="129086f62452bf149b797fc37d111e6a4234c3e372415e4ccc6270e25e57f59b" exitCode=0 Feb 22 00:09:30 crc kubenswrapper[4857]: I0222 00:09:30.585565 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29528640-g7sjv" event={"ID":"f07c5f6c-743d-4032-b70b-ecc0f514295d","Type":"ContainerDied","Data":"129086f62452bf149b797fc37d111e6a4234c3e372415e4ccc6270e25e57f59b"} Feb 22 00:09:34 crc kubenswrapper[4857]: I0222 00:09:34.945709 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2p58w" Feb 22 00:09:38 crc kubenswrapper[4857]: E0222 00:09:38.807721 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 22 00:09:38 crc kubenswrapper[4857]: E0222 00:09:38.808437 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k9fqw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-fmz46_openshift-marketplace(299c4a35-6470-4a40-9b90-ad0390497677): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 22 00:09:38 crc kubenswrapper[4857]: E0222 00:09:38.811228 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-fmz46" podUID="299c4a35-6470-4a40-9b90-ad0390497677" Feb 22 00:09:39 crc kubenswrapper[4857]: I0222 00:09:39.027863 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29528640-g7sjv" Feb 22 00:09:39 crc kubenswrapper[4857]: E0222 00:09:39.113760 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 22 00:09:39 crc kubenswrapper[4857]: E0222 00:09:39.113899 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cd2qp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-xd4g7_openshift-marketplace(6d73807b-5e0c-4607-8393-5123a6cdbf9b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 22 00:09:39 crc kubenswrapper[4857]: E0222 00:09:39.115307 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-xd4g7" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" Feb 22 00:09:39 crc kubenswrapper[4857]: I0222 00:09:39.213081 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s48vw\" (UniqueName: \"kubernetes.io/projected/f07c5f6c-743d-4032-b70b-ecc0f514295d-kube-api-access-s48vw\") pod \"f07c5f6c-743d-4032-b70b-ecc0f514295d\" (UID: \"f07c5f6c-743d-4032-b70b-ecc0f514295d\") " Feb 22 00:09:39 crc kubenswrapper[4857]: I0222 00:09:39.213555 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f07c5f6c-743d-4032-b70b-ecc0f514295d-serviceca\") pod \"f07c5f6c-743d-4032-b70b-ecc0f514295d\" (UID: \"f07c5f6c-743d-4032-b70b-ecc0f514295d\") " Feb 22 00:09:39 crc kubenswrapper[4857]: I0222 00:09:39.214527 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f07c5f6c-743d-4032-b70b-ecc0f514295d-serviceca" (OuterVolumeSpecName: "serviceca") pod "f07c5f6c-743d-4032-b70b-ecc0f514295d" (UID: "f07c5f6c-743d-4032-b70b-ecc0f514295d"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:09:39 crc kubenswrapper[4857]: I0222 00:09:39.225661 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f07c5f6c-743d-4032-b70b-ecc0f514295d-kube-api-access-s48vw" (OuterVolumeSpecName: "kube-api-access-s48vw") pod "f07c5f6c-743d-4032-b70b-ecc0f514295d" (UID: "f07c5f6c-743d-4032-b70b-ecc0f514295d"). InnerVolumeSpecName "kube-api-access-s48vw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:09:39 crc kubenswrapper[4857]: E0222 00:09:39.242363 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 22 00:09:39 crc kubenswrapper[4857]: E0222 00:09:39.242555 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ml52g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-bdmrq_openshift-marketplace(0c14592b-a2d9-46c6-b818-6da2942cb895): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 22 00:09:39 crc kubenswrapper[4857]: E0222 00:09:39.243716 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-bdmrq" podUID="0c14592b-a2d9-46c6-b818-6da2942cb895" Feb 22 00:09:39 crc kubenswrapper[4857]: I0222 00:09:39.315003 4857 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f07c5f6c-743d-4032-b70b-ecc0f514295d-serviceca\") on node \"crc\" DevicePath \"\"" Feb 22 00:09:39 crc kubenswrapper[4857]: I0222 00:09:39.315059 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s48vw\" (UniqueName: \"kubernetes.io/projected/f07c5f6c-743d-4032-b70b-ecc0f514295d-kube-api-access-s48vw\") on node \"crc\" DevicePath \"\"" Feb 22 00:09:39 crc kubenswrapper[4857]: I0222 00:09:39.488434 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-jnm7q"] Feb 22 00:09:39 crc kubenswrapper[4857]: I0222 00:09:39.642745 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-797dx" event={"ID":"10f2d337-12f1-4949-9872-08df9aa084ab","Type":"ContainerStarted","Data":"d2baba890384c306149473909fe2096c933348ac570a9a5e78479881563b25f3"} Feb 22 00:09:39 crc kubenswrapper[4857]: I0222 00:09:39.644348 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29528640-g7sjv" event={"ID":"f07c5f6c-743d-4032-b70b-ecc0f514295d","Type":"ContainerDied","Data":"4bcb8686a27b403e0c8cbe7bc18c2a7f298997eae2c838b76ca4e10000d0154e"} Feb 22 00:09:39 crc kubenswrapper[4857]: I0222 00:09:39.644368 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29528640-g7sjv" Feb 22 00:09:39 crc kubenswrapper[4857]: I0222 00:09:39.645090 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bcb8686a27b403e0c8cbe7bc18c2a7f298997eae2c838b76ca4e10000d0154e" Feb 22 00:09:39 crc kubenswrapper[4857]: I0222 00:09:39.646814 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jw5r7" event={"ID":"0ced68fc-a805-4595-8043-ce7ec0b802a3","Type":"ContainerStarted","Data":"2cc4ee7b1abd53d91db077ada2e54fd9288145e7ab3a6a4a313c6b348773aaf5"} Feb 22 00:09:39 crc kubenswrapper[4857]: I0222 00:09:39.648699 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5rft" event={"ID":"18611d84-22b3-49df-9569-ac383eb42cab","Type":"ContainerStarted","Data":"d807fcd19722a12043390537664dbf2f93cdad07a48817b2d3130ba55755dd6d"} Feb 22 00:09:39 crc kubenswrapper[4857]: I0222 00:09:39.650380 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rpdr" event={"ID":"fa45663f-d64c-431c-be1d-29175cc804ea","Type":"ContainerStarted","Data":"246cf699c95224893fcba082295e051b68ac17f47fec785c15e5741502bac507"} Feb 22 00:09:39 crc kubenswrapper[4857]: I0222 00:09:39.651547 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" event={"ID":"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53","Type":"ContainerStarted","Data":"f0f01d7639dacd6a6a6b650fb7868e7af44852b16ea4b7244db525dd9b276751"} Feb 22 00:09:39 crc kubenswrapper[4857]: I0222 00:09:39.653931 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqp5c" event={"ID":"d905b043-15bc-4147-8745-d7c26f087559","Type":"ContainerStarted","Data":"96e2b65cf07b200951d1b0e644196d07ef56d80dc72ba8e45eedca708656af42"} Feb 22 00:09:39 crc kubenswrapper[4857]: E0222 00:09:39.654787 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-xd4g7" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" Feb 22 00:09:39 crc kubenswrapper[4857]: E0222 00:09:39.657287 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-fmz46" podUID="299c4a35-6470-4a40-9b90-ad0390497677" Feb 22 00:09:39 crc kubenswrapper[4857]: E0222 00:09:39.658821 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-bdmrq" podUID="0c14592b-a2d9-46c6-b818-6da2942cb895" Feb 22 00:09:40 crc kubenswrapper[4857]: I0222 00:09:40.428396 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:09:40 crc kubenswrapper[4857]: I0222 00:09:40.428465 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:09:40 crc kubenswrapper[4857]: I0222 00:09:40.663874 4857 generic.go:334] "Generic (PLEG): container finished" podID="d905b043-15bc-4147-8745-d7c26f087559" containerID="96e2b65cf07b200951d1b0e644196d07ef56d80dc72ba8e45eedca708656af42" exitCode=0 Feb 22 00:09:40 crc kubenswrapper[4857]: I0222 00:09:40.663950 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqp5c" event={"ID":"d905b043-15bc-4147-8745-d7c26f087559","Type":"ContainerDied","Data":"96e2b65cf07b200951d1b0e644196d07ef56d80dc72ba8e45eedca708656af42"} Feb 22 00:09:40 crc kubenswrapper[4857]: I0222 00:09:40.667774 4857 generic.go:334] "Generic (PLEG): container finished" podID="10f2d337-12f1-4949-9872-08df9aa084ab" containerID="d2baba890384c306149473909fe2096c933348ac570a9a5e78479881563b25f3" exitCode=0 Feb 22 00:09:40 crc kubenswrapper[4857]: I0222 00:09:40.667848 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-797dx" event={"ID":"10f2d337-12f1-4949-9872-08df9aa084ab","Type":"ContainerDied","Data":"d2baba890384c306149473909fe2096c933348ac570a9a5e78479881563b25f3"} Feb 22 00:09:40 crc kubenswrapper[4857]: I0222 00:09:40.671074 4857 generic.go:334] "Generic (PLEG): container finished" podID="0ced68fc-a805-4595-8043-ce7ec0b802a3" containerID="2cc4ee7b1abd53d91db077ada2e54fd9288145e7ab3a6a4a313c6b348773aaf5" exitCode=0 Feb 22 00:09:40 crc kubenswrapper[4857]: I0222 00:09:40.671157 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jw5r7" event={"ID":"0ced68fc-a805-4595-8043-ce7ec0b802a3","Type":"ContainerDied","Data":"2cc4ee7b1abd53d91db077ada2e54fd9288145e7ab3a6a4a313c6b348773aaf5"} Feb 22 00:09:40 crc kubenswrapper[4857]: I0222 00:09:40.674437 4857 generic.go:334] "Generic (PLEG): container finished" podID="18611d84-22b3-49df-9569-ac383eb42cab" containerID="d807fcd19722a12043390537664dbf2f93cdad07a48817b2d3130ba55755dd6d" exitCode=0 Feb 22 00:09:40 crc kubenswrapper[4857]: I0222 00:09:40.674530 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5rft" event={"ID":"18611d84-22b3-49df-9569-ac383eb42cab","Type":"ContainerDied","Data":"d807fcd19722a12043390537664dbf2f93cdad07a48817b2d3130ba55755dd6d"} Feb 22 00:09:40 crc kubenswrapper[4857]: I0222 00:09:40.682481 4857 generic.go:334] "Generic (PLEG): container finished" podID="fa45663f-d64c-431c-be1d-29175cc804ea" containerID="246cf699c95224893fcba082295e051b68ac17f47fec785c15e5741502bac507" exitCode=0 Feb 22 00:09:40 crc kubenswrapper[4857]: I0222 00:09:40.682563 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rpdr" event={"ID":"fa45663f-d64c-431c-be1d-29175cc804ea","Type":"ContainerDied","Data":"246cf699c95224893fcba082295e051b68ac17f47fec785c15e5741502bac507"} Feb 22 00:09:40 crc kubenswrapper[4857]: I0222 00:09:40.684335 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" event={"ID":"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53","Type":"ContainerStarted","Data":"24d3282555a5d61f767249bcb02196d8ffe6ba688cc5e506fa26e00d69e7dfc0"} Feb 22 00:09:41 crc kubenswrapper[4857]: I0222 00:09:41.133654 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 22 00:09:41 crc kubenswrapper[4857]: I0222 00:09:41.690678 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jnm7q" event={"ID":"a0fb0fe8-ad7a-48a5-8a76-737cd5076f53","Type":"ContainerStarted","Data":"56172cca0ba91c89922b064cb761d48139cb41f2745ce071320d6da3f8413ab9"} Feb 22 00:09:42 crc kubenswrapper[4857]: I0222 00:09:42.557605 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-jnm7q" podStartSLOduration=166.557581575 podStartE2EDuration="2m46.557581575s" podCreationTimestamp="2026-02-22 00:06:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:09:41.710848995 +0000 UTC m=+189.349578248" watchObservedRunningTime="2026-02-22 00:09:42.557581575 +0000 UTC m=+190.196310838" Feb 22 00:09:42 crc kubenswrapper[4857]: I0222 00:09:42.560449 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lm22m"] Feb 22 00:09:42 crc kubenswrapper[4857]: I0222 00:09:42.696709 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqp5c" event={"ID":"d905b043-15bc-4147-8745-d7c26f087559","Type":"ContainerStarted","Data":"15379592055c0a839f218d9d4ea485d8e159322b7c84f1aa9994d374520f273b"} Feb 22 00:09:42 crc kubenswrapper[4857]: I0222 00:09:42.698542 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jw5r7" event={"ID":"0ced68fc-a805-4595-8043-ce7ec0b802a3","Type":"ContainerStarted","Data":"e55fdcc8cc9a59f05cebb1a9fa3ce1bcbef8532cfe1975046947b5c75cc59dc4"} Feb 22 00:09:42 crc kubenswrapper[4857]: I0222 00:09:42.700394 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5rft" event={"ID":"18611d84-22b3-49df-9569-ac383eb42cab","Type":"ContainerStarted","Data":"7ba7afef408b9c9fe34536b3e36156921afe7f08cf8f984219ea7dc53fbd01cf"} Feb 22 00:09:42 crc kubenswrapper[4857]: I0222 00:09:42.702647 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rpdr" event={"ID":"fa45663f-d64c-431c-be1d-29175cc804ea","Type":"ContainerStarted","Data":"0d3f598c3d03bf8ce8f4bef7c23fdabe3795b3736cc7a62fc550d7a372dca74c"} Feb 22 00:09:42 crc kubenswrapper[4857]: I0222 00:09:42.722722 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nqp5c" podStartSLOduration=2.653935555 podStartE2EDuration="40.722704293s" podCreationTimestamp="2026-02-22 00:09:02 +0000 UTC" firstStartedPulling="2026-02-22 00:09:04.239670414 +0000 UTC m=+151.878399667" lastFinishedPulling="2026-02-22 00:09:42.308439152 +0000 UTC m=+189.947168405" observedRunningTime="2026-02-22 00:09:42.72154347 +0000 UTC m=+190.360272733" watchObservedRunningTime="2026-02-22 00:09:42.722704293 +0000 UTC m=+190.361433546" Feb 22 00:09:42 crc kubenswrapper[4857]: I0222 00:09:42.747005 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5rpdr" podStartSLOduration=1.713869323 podStartE2EDuration="40.746988916s" podCreationTimestamp="2026-02-22 00:09:02 +0000 UTC" firstStartedPulling="2026-02-22 00:09:03.196474482 +0000 UTC m=+150.835203745" lastFinishedPulling="2026-02-22 00:09:42.229594095 +0000 UTC m=+189.868323338" observedRunningTime="2026-02-22 00:09:42.744547266 +0000 UTC m=+190.383276519" watchObservedRunningTime="2026-02-22 00:09:42.746988916 +0000 UTC m=+190.385718169" Feb 22 00:09:42 crc kubenswrapper[4857]: I0222 00:09:42.766478 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jw5r7" podStartSLOduration=3.2843926789999998 podStartE2EDuration="40.766463171s" podCreationTimestamp="2026-02-22 00:09:02 +0000 UTC" firstStartedPulling="2026-02-22 00:09:04.215120764 +0000 UTC m=+151.853850017" lastFinishedPulling="2026-02-22 00:09:41.697191256 +0000 UTC m=+189.335920509" observedRunningTime="2026-02-22 00:09:42.763453945 +0000 UTC m=+190.402183208" watchObservedRunningTime="2026-02-22 00:09:42.766463171 +0000 UTC m=+190.405192424" Feb 22 00:09:42 crc kubenswrapper[4857]: I0222 00:09:42.793152 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jw5r7" Feb 22 00:09:42 crc kubenswrapper[4857]: I0222 00:09:42.793203 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jw5r7" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.019510 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m5rft" podStartSLOduration=2.8255928580000003 podStartE2EDuration="41.019488624s" podCreationTimestamp="2026-02-22 00:09:02 +0000 UTC" firstStartedPulling="2026-02-22 00:09:04.218285644 +0000 UTC m=+151.857014897" lastFinishedPulling="2026-02-22 00:09:42.41218141 +0000 UTC m=+190.050910663" observedRunningTime="2026-02-22 00:09:42.791013691 +0000 UTC m=+190.429742944" watchObservedRunningTime="2026-02-22 00:09:43.019488624 +0000 UTC m=+190.658217887" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.022286 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 22 00:09:43 crc kubenswrapper[4857]: E0222 00:09:43.022549 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feceeec5-56fd-4977-9f97-f0dc03546b98" containerName="pruner" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.022571 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="feceeec5-56fd-4977-9f97-f0dc03546b98" containerName="pruner" Feb 22 00:09:43 crc kubenswrapper[4857]: E0222 00:09:43.022588 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f07c5f6c-743d-4032-b70b-ecc0f514295d" containerName="image-pruner" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.022597 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f07c5f6c-743d-4032-b70b-ecc0f514295d" containerName="image-pruner" Feb 22 00:09:43 crc kubenswrapper[4857]: E0222 00:09:43.022613 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="112060a6-9baa-411b-9e8f-9f57c271a1d9" containerName="pruner" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.022621 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="112060a6-9baa-411b-9e8f-9f57c271a1d9" containerName="pruner" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.022735 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="112060a6-9baa-411b-9e8f-9f57c271a1d9" containerName="pruner" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.022753 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f07c5f6c-743d-4032-b70b-ecc0f514295d" containerName="image-pruner" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.022764 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="feceeec5-56fd-4977-9f97-f0dc03546b98" containerName="pruner" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.023250 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.026230 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.026448 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.034654 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.054236 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nqp5c" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.054286 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nqp5c" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.169283 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0d90eed-681f-49af-a887-52bc2ff4199f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a0d90eed-681f-49af-a887-52bc2ff4199f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.169587 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0d90eed-681f-49af-a887-52bc2ff4199f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a0d90eed-681f-49af-a887-52bc2ff4199f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.270776 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0d90eed-681f-49af-a887-52bc2ff4199f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a0d90eed-681f-49af-a887-52bc2ff4199f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.270842 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0d90eed-681f-49af-a887-52bc2ff4199f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a0d90eed-681f-49af-a887-52bc2ff4199f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.271242 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0d90eed-681f-49af-a887-52bc2ff4199f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a0d90eed-681f-49af-a887-52bc2ff4199f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.288585 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0d90eed-681f-49af-a887-52bc2ff4199f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a0d90eed-681f-49af-a887-52bc2ff4199f\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.339642 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.583047 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 22 00:09:43 crc kubenswrapper[4857]: W0222 00:09:43.590907 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda0d90eed_681f_49af_a887_52bc2ff4199f.slice/crio-016f8bac9ebf9d3208ba856c90266d2e5d7f8c882ca5ad8a0da4546b2a8753e5 WatchSource:0}: Error finding container 016f8bac9ebf9d3208ba856c90266d2e5d7f8c882ca5ad8a0da4546b2a8753e5: Status 404 returned error can't find the container with id 016f8bac9ebf9d3208ba856c90266d2e5d7f8c882ca5ad8a0da4546b2a8753e5 Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.717972 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-797dx" event={"ID":"10f2d337-12f1-4949-9872-08df9aa084ab","Type":"ContainerStarted","Data":"5d7a76f88ed88af6871dfb7951a544360508850bbf76aab186a86107c836c6fb"} Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.723647 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a0d90eed-681f-49af-a887-52bc2ff4199f","Type":"ContainerStarted","Data":"016f8bac9ebf9d3208ba856c90266d2e5d7f8c882ca5ad8a0da4546b2a8753e5"} Feb 22 00:09:43 crc kubenswrapper[4857]: I0222 00:09:43.937321 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-jw5r7" podUID="0ced68fc-a805-4595-8043-ce7ec0b802a3" containerName="registry-server" probeResult="failure" output=< Feb 22 00:09:43 crc kubenswrapper[4857]: timeout: failed to connect service ":50051" within 1s Feb 22 00:09:43 crc kubenswrapper[4857]: > Feb 22 00:09:44 crc kubenswrapper[4857]: I0222 00:09:44.100926 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-nqp5c" podUID="d905b043-15bc-4147-8745-d7c26f087559" containerName="registry-server" probeResult="failure" output=< Feb 22 00:09:44 crc kubenswrapper[4857]: timeout: failed to connect service ":50051" within 1s Feb 22 00:09:44 crc kubenswrapper[4857]: > Feb 22 00:09:44 crc kubenswrapper[4857]: I0222 00:09:44.618249 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-797dx" Feb 22 00:09:44 crc kubenswrapper[4857]: I0222 00:09:44.618351 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-797dx" Feb 22 00:09:44 crc kubenswrapper[4857]: I0222 00:09:44.730670 4857 generic.go:334] "Generic (PLEG): container finished" podID="a0d90eed-681f-49af-a887-52bc2ff4199f" containerID="eeca049f8c537c523d4ec60f83e0a283de6387128aa719e56207951d6b1ee197" exitCode=0 Feb 22 00:09:44 crc kubenswrapper[4857]: I0222 00:09:44.730729 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a0d90eed-681f-49af-a887-52bc2ff4199f","Type":"ContainerDied","Data":"eeca049f8c537c523d4ec60f83e0a283de6387128aa719e56207951d6b1ee197"} Feb 22 00:09:44 crc kubenswrapper[4857]: I0222 00:09:44.742868 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-797dx" podStartSLOduration=3.3934156189999998 podStartE2EDuration="40.742849199s" podCreationTimestamp="2026-02-22 00:09:04 +0000 UTC" firstStartedPulling="2026-02-22 00:09:05.23564766 +0000 UTC m=+152.874376913" lastFinishedPulling="2026-02-22 00:09:42.58508124 +0000 UTC m=+190.223810493" observedRunningTime="2026-02-22 00:09:43.737706292 +0000 UTC m=+191.376435555" watchObservedRunningTime="2026-02-22 00:09:44.742849199 +0000 UTC m=+192.381578452" Feb 22 00:09:45 crc kubenswrapper[4857]: I0222 00:09:45.666782 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-797dx" podUID="10f2d337-12f1-4949-9872-08df9aa084ab" containerName="registry-server" probeResult="failure" output=< Feb 22 00:09:45 crc kubenswrapper[4857]: timeout: failed to connect service ":50051" within 1s Feb 22 00:09:45 crc kubenswrapper[4857]: > Feb 22 00:09:45 crc kubenswrapper[4857]: I0222 00:09:45.919225 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 22 00:09:46 crc kubenswrapper[4857]: I0222 00:09:46.108632 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0d90eed-681f-49af-a887-52bc2ff4199f-kube-api-access\") pod \"a0d90eed-681f-49af-a887-52bc2ff4199f\" (UID: \"a0d90eed-681f-49af-a887-52bc2ff4199f\") " Feb 22 00:09:46 crc kubenswrapper[4857]: I0222 00:09:46.108700 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0d90eed-681f-49af-a887-52bc2ff4199f-kubelet-dir\") pod \"a0d90eed-681f-49af-a887-52bc2ff4199f\" (UID: \"a0d90eed-681f-49af-a887-52bc2ff4199f\") " Feb 22 00:09:46 crc kubenswrapper[4857]: I0222 00:09:46.108964 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0d90eed-681f-49af-a887-52bc2ff4199f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a0d90eed-681f-49af-a887-52bc2ff4199f" (UID: "a0d90eed-681f-49af-a887-52bc2ff4199f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:09:46 crc kubenswrapper[4857]: I0222 00:09:46.118235 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0d90eed-681f-49af-a887-52bc2ff4199f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a0d90eed-681f-49af-a887-52bc2ff4199f" (UID: "a0d90eed-681f-49af-a887-52bc2ff4199f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:09:46 crc kubenswrapper[4857]: I0222 00:09:46.209891 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a0d90eed-681f-49af-a887-52bc2ff4199f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 22 00:09:46 crc kubenswrapper[4857]: I0222 00:09:46.210206 4857 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a0d90eed-681f-49af-a887-52bc2ff4199f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 22 00:09:46 crc kubenswrapper[4857]: I0222 00:09:46.741398 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a0d90eed-681f-49af-a887-52bc2ff4199f","Type":"ContainerDied","Data":"016f8bac9ebf9d3208ba856c90266d2e5d7f8c882ca5ad8a0da4546b2a8753e5"} Feb 22 00:09:46 crc kubenswrapper[4857]: I0222 00:09:46.741442 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="016f8bac9ebf9d3208ba856c90266d2e5d7f8c882ca5ad8a0da4546b2a8753e5" Feb 22 00:09:46 crc kubenswrapper[4857]: I0222 00:09:46.741461 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 22 00:09:50 crc kubenswrapper[4857]: I0222 00:09:50.219925 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 22 00:09:50 crc kubenswrapper[4857]: E0222 00:09:50.220189 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0d90eed-681f-49af-a887-52bc2ff4199f" containerName="pruner" Feb 22 00:09:50 crc kubenswrapper[4857]: I0222 00:09:50.220200 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0d90eed-681f-49af-a887-52bc2ff4199f" containerName="pruner" Feb 22 00:09:50 crc kubenswrapper[4857]: I0222 00:09:50.220293 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0d90eed-681f-49af-a887-52bc2ff4199f" containerName="pruner" Feb 22 00:09:50 crc kubenswrapper[4857]: I0222 00:09:50.220652 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 22 00:09:50 crc kubenswrapper[4857]: I0222 00:09:50.222803 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 22 00:09:50 crc kubenswrapper[4857]: I0222 00:09:50.223280 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 22 00:09:50 crc kubenswrapper[4857]: I0222 00:09:50.231598 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 22 00:09:50 crc kubenswrapper[4857]: I0222 00:09:50.358420 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7ab5d710-d809-4d86-b56b-1f611cb02644-var-lock\") pod \"installer-9-crc\" (UID: \"7ab5d710-d809-4d86-b56b-1f611cb02644\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 22 00:09:50 crc kubenswrapper[4857]: I0222 00:09:50.358467 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ab5d710-d809-4d86-b56b-1f611cb02644-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7ab5d710-d809-4d86-b56b-1f611cb02644\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 22 00:09:50 crc kubenswrapper[4857]: I0222 00:09:50.358491 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ab5d710-d809-4d86-b56b-1f611cb02644-kube-api-access\") pod \"installer-9-crc\" (UID: \"7ab5d710-d809-4d86-b56b-1f611cb02644\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 22 00:09:50 crc kubenswrapper[4857]: I0222 00:09:50.459536 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7ab5d710-d809-4d86-b56b-1f611cb02644-var-lock\") pod \"installer-9-crc\" (UID: \"7ab5d710-d809-4d86-b56b-1f611cb02644\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 22 00:09:50 crc kubenswrapper[4857]: I0222 00:09:50.459586 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ab5d710-d809-4d86-b56b-1f611cb02644-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7ab5d710-d809-4d86-b56b-1f611cb02644\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 22 00:09:50 crc kubenswrapper[4857]: I0222 00:09:50.459611 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ab5d710-d809-4d86-b56b-1f611cb02644-kube-api-access\") pod \"installer-9-crc\" (UID: \"7ab5d710-d809-4d86-b56b-1f611cb02644\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 22 00:09:50 crc kubenswrapper[4857]: I0222 00:09:50.459686 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ab5d710-d809-4d86-b56b-1f611cb02644-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7ab5d710-d809-4d86-b56b-1f611cb02644\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 22 00:09:50 crc kubenswrapper[4857]: I0222 00:09:50.459722 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7ab5d710-d809-4d86-b56b-1f611cb02644-var-lock\") pod \"installer-9-crc\" (UID: \"7ab5d710-d809-4d86-b56b-1f611cb02644\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 22 00:09:50 crc kubenswrapper[4857]: I0222 00:09:50.480817 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ab5d710-d809-4d86-b56b-1f611cb02644-kube-api-access\") pod \"installer-9-crc\" (UID: \"7ab5d710-d809-4d86-b56b-1f611cb02644\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 22 00:09:50 crc kubenswrapper[4857]: I0222 00:09:50.551962 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 22 00:09:50 crc kubenswrapper[4857]: I0222 00:09:50.775835 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 22 00:09:50 crc kubenswrapper[4857]: W0222 00:09:50.784211 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7ab5d710_d809_4d86_b56b_1f611cb02644.slice/crio-7bf51deb2895b48144471757f30daa5e82610e4d537ea4fa71ae9d2edda4f000 WatchSource:0}: Error finding container 7bf51deb2895b48144471757f30daa5e82610e4d537ea4fa71ae9d2edda4f000: Status 404 returned error can't find the container with id 7bf51deb2895b48144471757f30daa5e82610e4d537ea4fa71ae9d2edda4f000 Feb 22 00:09:51 crc kubenswrapper[4857]: I0222 00:09:51.767845 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7ab5d710-d809-4d86-b56b-1f611cb02644","Type":"ContainerStarted","Data":"7bf51deb2895b48144471757f30daa5e82610e4d537ea4fa71ae9d2edda4f000"} Feb 22 00:09:52 crc kubenswrapper[4857]: I0222 00:09:52.402243 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5rpdr" Feb 22 00:09:52 crc kubenswrapper[4857]: I0222 00:09:52.402273 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5rpdr" Feb 22 00:09:52 crc kubenswrapper[4857]: I0222 00:09:52.463880 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5rpdr" Feb 22 00:09:52 crc kubenswrapper[4857]: I0222 00:09:52.615335 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m5rft" Feb 22 00:09:52 crc kubenswrapper[4857]: I0222 00:09:52.616165 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m5rft" Feb 22 00:09:52 crc kubenswrapper[4857]: I0222 00:09:52.653607 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m5rft" Feb 22 00:09:52 crc kubenswrapper[4857]: I0222 00:09:52.774208 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7ab5d710-d809-4d86-b56b-1f611cb02644","Type":"ContainerStarted","Data":"3694ae29f8d75fe25793827365d4a12165d51adc31c47f3273747ed2a84bdfa4"} Feb 22 00:09:52 crc kubenswrapper[4857]: I0222 00:09:52.808558 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5rpdr" Feb 22 00:09:52 crc kubenswrapper[4857]: I0222 00:09:52.809096 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m5rft" Feb 22 00:09:52 crc kubenswrapper[4857]: I0222 00:09:52.845015 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jw5r7" Feb 22 00:09:52 crc kubenswrapper[4857]: I0222 00:09:52.882939 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jw5r7" Feb 22 00:09:53 crc kubenswrapper[4857]: I0222 00:09:53.099708 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nqp5c" Feb 22 00:09:53 crc kubenswrapper[4857]: I0222 00:09:53.144133 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nqp5c" Feb 22 00:09:53 crc kubenswrapper[4857]: I0222 00:09:53.708312 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nqp5c"] Feb 22 00:09:53 crc kubenswrapper[4857]: I0222 00:09:53.794425 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=3.794408352 podStartE2EDuration="3.794408352s" podCreationTimestamp="2026-02-22 00:09:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:09:53.793250122 +0000 UTC m=+201.431979385" watchObservedRunningTime="2026-02-22 00:09:53.794408352 +0000 UTC m=+201.433137625" Feb 22 00:09:54 crc kubenswrapper[4857]: I0222 00:09:54.654532 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-797dx" Feb 22 00:09:54 crc kubenswrapper[4857]: I0222 00:09:54.688915 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-797dx" Feb 22 00:09:54 crc kubenswrapper[4857]: I0222 00:09:54.789456 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nqp5c" podUID="d905b043-15bc-4147-8745-d7c26f087559" containerName="registry-server" containerID="cri-o://15379592055c0a839f218d9d4ea485d8e159322b7c84f1aa9994d374520f273b" gracePeriod=2 Feb 22 00:09:55 crc kubenswrapper[4857]: I0222 00:09:55.107396 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jw5r7"] Feb 22 00:09:55 crc kubenswrapper[4857]: I0222 00:09:55.107951 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jw5r7" podUID="0ced68fc-a805-4595-8043-ce7ec0b802a3" containerName="registry-server" containerID="cri-o://e55fdcc8cc9a59f05cebb1a9fa3ce1bcbef8532cfe1975046947b5c75cc59dc4" gracePeriod=2 Feb 22 00:09:55 crc kubenswrapper[4857]: I0222 00:09:55.797730 4857 generic.go:334] "Generic (PLEG): container finished" podID="d905b043-15bc-4147-8745-d7c26f087559" containerID="15379592055c0a839f218d9d4ea485d8e159322b7c84f1aa9994d374520f273b" exitCode=0 Feb 22 00:09:55 crc kubenswrapper[4857]: I0222 00:09:55.797778 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqp5c" event={"ID":"d905b043-15bc-4147-8745-d7c26f087559","Type":"ContainerDied","Data":"15379592055c0a839f218d9d4ea485d8e159322b7c84f1aa9994d374520f273b"} Feb 22 00:09:56 crc kubenswrapper[4857]: I0222 00:09:56.805454 4857 generic.go:334] "Generic (PLEG): container finished" podID="0ced68fc-a805-4595-8043-ce7ec0b802a3" containerID="e55fdcc8cc9a59f05cebb1a9fa3ce1bcbef8532cfe1975046947b5c75cc59dc4" exitCode=0 Feb 22 00:09:56 crc kubenswrapper[4857]: I0222 00:09:56.805506 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jw5r7" event={"ID":"0ced68fc-a805-4595-8043-ce7ec0b802a3","Type":"ContainerDied","Data":"e55fdcc8cc9a59f05cebb1a9fa3ce1bcbef8532cfe1975046947b5c75cc59dc4"} Feb 22 00:09:57 crc kubenswrapper[4857]: I0222 00:09:57.999520 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jw5r7" Feb 22 00:09:58 crc kubenswrapper[4857]: I0222 00:09:58.065828 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ced68fc-a805-4595-8043-ce7ec0b802a3-catalog-content\") pod \"0ced68fc-a805-4595-8043-ce7ec0b802a3\" (UID: \"0ced68fc-a805-4595-8043-ce7ec0b802a3\") " Feb 22 00:09:58 crc kubenswrapper[4857]: I0222 00:09:58.065875 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rc2j\" (UniqueName: \"kubernetes.io/projected/0ced68fc-a805-4595-8043-ce7ec0b802a3-kube-api-access-5rc2j\") pod \"0ced68fc-a805-4595-8043-ce7ec0b802a3\" (UID: \"0ced68fc-a805-4595-8043-ce7ec0b802a3\") " Feb 22 00:09:58 crc kubenswrapper[4857]: I0222 00:09:58.065937 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ced68fc-a805-4595-8043-ce7ec0b802a3-utilities\") pod \"0ced68fc-a805-4595-8043-ce7ec0b802a3\" (UID: \"0ced68fc-a805-4595-8043-ce7ec0b802a3\") " Feb 22 00:09:58 crc kubenswrapper[4857]: I0222 00:09:58.066927 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ced68fc-a805-4595-8043-ce7ec0b802a3-utilities" (OuterVolumeSpecName: "utilities") pod "0ced68fc-a805-4595-8043-ce7ec0b802a3" (UID: "0ced68fc-a805-4595-8043-ce7ec0b802a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:09:58 crc kubenswrapper[4857]: I0222 00:09:58.075344 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ced68fc-a805-4595-8043-ce7ec0b802a3-kube-api-access-5rc2j" (OuterVolumeSpecName: "kube-api-access-5rc2j") pod "0ced68fc-a805-4595-8043-ce7ec0b802a3" (UID: "0ced68fc-a805-4595-8043-ce7ec0b802a3"). InnerVolumeSpecName "kube-api-access-5rc2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:09:58 crc kubenswrapper[4857]: I0222 00:09:58.111881 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ced68fc-a805-4595-8043-ce7ec0b802a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ced68fc-a805-4595-8043-ce7ec0b802a3" (UID: "0ced68fc-a805-4595-8043-ce7ec0b802a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:09:58 crc kubenswrapper[4857]: I0222 00:09:58.167738 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ced68fc-a805-4595-8043-ce7ec0b802a3-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:09:58 crc kubenswrapper[4857]: I0222 00:09:58.167771 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ced68fc-a805-4595-8043-ce7ec0b802a3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:09:58 crc kubenswrapper[4857]: I0222 00:09:58.167783 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rc2j\" (UniqueName: \"kubernetes.io/projected/0ced68fc-a805-4595-8043-ce7ec0b802a3-kube-api-access-5rc2j\") on node \"crc\" DevicePath \"\"" Feb 22 00:09:58 crc kubenswrapper[4857]: I0222 00:09:58.818299 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jw5r7" event={"ID":"0ced68fc-a805-4595-8043-ce7ec0b802a3","Type":"ContainerDied","Data":"c3bce4f9f27aae2e65ffc5b59da7d061629a0e0a91b9cead4d28f75a73e2e23c"} Feb 22 00:09:58 crc kubenswrapper[4857]: I0222 00:09:58.818346 4857 scope.go:117] "RemoveContainer" containerID="e55fdcc8cc9a59f05cebb1a9fa3ce1bcbef8532cfe1975046947b5c75cc59dc4" Feb 22 00:09:58 crc kubenswrapper[4857]: I0222 00:09:58.818443 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jw5r7" Feb 22 00:09:58 crc kubenswrapper[4857]: I0222 00:09:58.847543 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jw5r7"] Feb 22 00:09:58 crc kubenswrapper[4857]: I0222 00:09:58.850808 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jw5r7"] Feb 22 00:09:59 crc kubenswrapper[4857]: I0222 00:09:59.084276 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ced68fc-a805-4595-8043-ce7ec0b802a3" path="/var/lib/kubelet/pods/0ced68fc-a805-4595-8043-ce7ec0b802a3/volumes" Feb 22 00:09:59 crc kubenswrapper[4857]: I0222 00:09:59.287063 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqp5c" Feb 22 00:09:59 crc kubenswrapper[4857]: I0222 00:09:59.378665 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d905b043-15bc-4147-8745-d7c26f087559-utilities\") pod \"d905b043-15bc-4147-8745-d7c26f087559\" (UID: \"d905b043-15bc-4147-8745-d7c26f087559\") " Feb 22 00:09:59 crc kubenswrapper[4857]: I0222 00:09:59.378794 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjn4d\" (UniqueName: \"kubernetes.io/projected/d905b043-15bc-4147-8745-d7c26f087559-kube-api-access-pjn4d\") pod \"d905b043-15bc-4147-8745-d7c26f087559\" (UID: \"d905b043-15bc-4147-8745-d7c26f087559\") " Feb 22 00:09:59 crc kubenswrapper[4857]: I0222 00:09:59.378852 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d905b043-15bc-4147-8745-d7c26f087559-catalog-content\") pod \"d905b043-15bc-4147-8745-d7c26f087559\" (UID: \"d905b043-15bc-4147-8745-d7c26f087559\") " Feb 22 00:09:59 crc kubenswrapper[4857]: I0222 00:09:59.380195 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d905b043-15bc-4147-8745-d7c26f087559-utilities" (OuterVolumeSpecName: "utilities") pod "d905b043-15bc-4147-8745-d7c26f087559" (UID: "d905b043-15bc-4147-8745-d7c26f087559"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:09:59 crc kubenswrapper[4857]: I0222 00:09:59.395658 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d905b043-15bc-4147-8745-d7c26f087559-kube-api-access-pjn4d" (OuterVolumeSpecName: "kube-api-access-pjn4d") pod "d905b043-15bc-4147-8745-d7c26f087559" (UID: "d905b043-15bc-4147-8745-d7c26f087559"). InnerVolumeSpecName "kube-api-access-pjn4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:09:59 crc kubenswrapper[4857]: I0222 00:09:59.432633 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d905b043-15bc-4147-8745-d7c26f087559-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d905b043-15bc-4147-8745-d7c26f087559" (UID: "d905b043-15bc-4147-8745-d7c26f087559"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:09:59 crc kubenswrapper[4857]: I0222 00:09:59.480300 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d905b043-15bc-4147-8745-d7c26f087559-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:09:59 crc kubenswrapper[4857]: I0222 00:09:59.480366 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjn4d\" (UniqueName: \"kubernetes.io/projected/d905b043-15bc-4147-8745-d7c26f087559-kube-api-access-pjn4d\") on node \"crc\" DevicePath \"\"" Feb 22 00:09:59 crc kubenswrapper[4857]: I0222 00:09:59.480395 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d905b043-15bc-4147-8745-d7c26f087559-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:09:59 crc kubenswrapper[4857]: I0222 00:09:59.826252 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nqp5c" event={"ID":"d905b043-15bc-4147-8745-d7c26f087559","Type":"ContainerDied","Data":"0d86d664b271e284c5aaf5f21b227c181bdaf0670ccbee8904682752eb20bc67"} Feb 22 00:09:59 crc kubenswrapper[4857]: I0222 00:09:59.826385 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nqp5c" Feb 22 00:09:59 crc kubenswrapper[4857]: I0222 00:09:59.856472 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nqp5c"] Feb 22 00:09:59 crc kubenswrapper[4857]: I0222 00:09:59.860930 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nqp5c"] Feb 22 00:10:00 crc kubenswrapper[4857]: I0222 00:10:00.204502 4857 scope.go:117] "RemoveContainer" containerID="2cc4ee7b1abd53d91db077ada2e54fd9288145e7ab3a6a4a313c6b348773aaf5" Feb 22 00:10:00 crc kubenswrapper[4857]: I0222 00:10:00.263480 4857 scope.go:117] "RemoveContainer" containerID="13ccc6e155863d7c55fdbe57ee7e38a666c9aba35c96187baa7d457023787ffa" Feb 22 00:10:00 crc kubenswrapper[4857]: I0222 00:10:00.338856 4857 scope.go:117] "RemoveContainer" containerID="15379592055c0a839f218d9d4ea485d8e159322b7c84f1aa9994d374520f273b" Feb 22 00:10:00 crc kubenswrapper[4857]: I0222 00:10:00.455324 4857 scope.go:117] "RemoveContainer" containerID="96e2b65cf07b200951d1b0e644196d07ef56d80dc72ba8e45eedca708656af42" Feb 22 00:10:00 crc kubenswrapper[4857]: I0222 00:10:00.538313 4857 scope.go:117] "RemoveContainer" containerID="2ca904adbfa7bfd8a4cc01573aaf71122c1db1d1c068aa74d09bcebac7377ead" Feb 22 00:10:01 crc kubenswrapper[4857]: I0222 00:10:01.083925 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d905b043-15bc-4147-8745-d7c26f087559" path="/var/lib/kubelet/pods/d905b043-15bc-4147-8745-d7c26f087559/volumes" Feb 22 00:10:01 crc kubenswrapper[4857]: I0222 00:10:01.840274 4857 generic.go:334] "Generic (PLEG): container finished" podID="0c14592b-a2d9-46c6-b818-6da2942cb895" containerID="c5065ca9d55267625ff6c51cfb84762429ca0f4be07e0ba01e2c413ba15f315b" exitCode=0 Feb 22 00:10:01 crc kubenswrapper[4857]: I0222 00:10:01.840304 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bdmrq" event={"ID":"0c14592b-a2d9-46c6-b818-6da2942cb895","Type":"ContainerDied","Data":"c5065ca9d55267625ff6c51cfb84762429ca0f4be07e0ba01e2c413ba15f315b"} Feb 22 00:10:01 crc kubenswrapper[4857]: I0222 00:10:01.843739 4857 generic.go:334] "Generic (PLEG): container finished" podID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" containerID="c75193c95e07cc6573aa6fd4397e64e48dba1f87c3b51cf693866eb540a1f56a" exitCode=0 Feb 22 00:10:01 crc kubenswrapper[4857]: I0222 00:10:01.843822 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xd4g7" event={"ID":"6d73807b-5e0c-4607-8393-5123a6cdbf9b","Type":"ContainerDied","Data":"c75193c95e07cc6573aa6fd4397e64e48dba1f87c3b51cf693866eb540a1f56a"} Feb 22 00:10:01 crc kubenswrapper[4857]: I0222 00:10:01.846717 4857 generic.go:334] "Generic (PLEG): container finished" podID="299c4a35-6470-4a40-9b90-ad0390497677" containerID="971edb974894b8c9cbd48f55ca1d5493160884bc60c723186931b064bae04712" exitCode=0 Feb 22 00:10:01 crc kubenswrapper[4857]: I0222 00:10:01.846757 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmz46" event={"ID":"299c4a35-6470-4a40-9b90-ad0390497677","Type":"ContainerDied","Data":"971edb974894b8c9cbd48f55ca1d5493160884bc60c723186931b064bae04712"} Feb 22 00:10:02 crc kubenswrapper[4857]: I0222 00:10:02.854202 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmz46" event={"ID":"299c4a35-6470-4a40-9b90-ad0390497677","Type":"ContainerStarted","Data":"7b5f75d1abc51d34be742c27a04c0b82bcf54997be03024b6f687faf71696113"} Feb 22 00:10:02 crc kubenswrapper[4857]: I0222 00:10:02.856779 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bdmrq" event={"ID":"0c14592b-a2d9-46c6-b818-6da2942cb895","Type":"ContainerStarted","Data":"f07bd44f1e70a4da4e3ad7e97db6f0ba6754e089402bcc5d506319adeed9d213"} Feb 22 00:10:02 crc kubenswrapper[4857]: I0222 00:10:02.859117 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xd4g7" event={"ID":"6d73807b-5e0c-4607-8393-5123a6cdbf9b","Type":"ContainerStarted","Data":"34c2234b5cda4e087d988f3152f4a10fc71a296bec05bfffd187d84e66ce1cf5"} Feb 22 00:10:02 crc kubenswrapper[4857]: I0222 00:10:02.873123 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fmz46" podStartSLOduration=2.935414262 podStartE2EDuration="57.873101761s" podCreationTimestamp="2026-02-22 00:09:05 +0000 UTC" firstStartedPulling="2026-02-22 00:09:07.390509425 +0000 UTC m=+155.029238678" lastFinishedPulling="2026-02-22 00:10:02.328196924 +0000 UTC m=+209.966926177" observedRunningTime="2026-02-22 00:10:02.872339718 +0000 UTC m=+210.511068981" watchObservedRunningTime="2026-02-22 00:10:02.873101761 +0000 UTC m=+210.511831024" Feb 22 00:10:02 crc kubenswrapper[4857]: I0222 00:10:02.897857 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bdmrq" podStartSLOduration=2.920051146 podStartE2EDuration="58.89783821s" podCreationTimestamp="2026-02-22 00:09:04 +0000 UTC" firstStartedPulling="2026-02-22 00:09:06.295428474 +0000 UTC m=+153.934157727" lastFinishedPulling="2026-02-22 00:10:02.273215538 +0000 UTC m=+209.911944791" observedRunningTime="2026-02-22 00:10:02.896511113 +0000 UTC m=+210.535240386" watchObservedRunningTime="2026-02-22 00:10:02.89783821 +0000 UTC m=+210.536567463" Feb 22 00:10:02 crc kubenswrapper[4857]: I0222 00:10:02.916337 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xd4g7" podStartSLOduration=3.06034249 podStartE2EDuration="57.916321422s" podCreationTimestamp="2026-02-22 00:09:05 +0000 UTC" firstStartedPulling="2026-02-22 00:09:07.379602814 +0000 UTC m=+155.018332067" lastFinishedPulling="2026-02-22 00:10:02.235581746 +0000 UTC m=+209.874310999" observedRunningTime="2026-02-22 00:10:02.911993187 +0000 UTC m=+210.550722440" watchObservedRunningTime="2026-02-22 00:10:02.916321422 +0000 UTC m=+210.555050675" Feb 22 00:10:04 crc kubenswrapper[4857]: I0222 00:10:04.996931 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bdmrq" Feb 22 00:10:04 crc kubenswrapper[4857]: I0222 00:10:04.997324 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bdmrq" Feb 22 00:10:05 crc kubenswrapper[4857]: I0222 00:10:05.037247 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bdmrq" Feb 22 00:10:05 crc kubenswrapper[4857]: I0222 00:10:05.649784 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xd4g7" Feb 22 00:10:05 crc kubenswrapper[4857]: I0222 00:10:05.650604 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xd4g7" Feb 22 00:10:06 crc kubenswrapper[4857]: I0222 00:10:06.022488 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fmz46" Feb 22 00:10:06 crc kubenswrapper[4857]: I0222 00:10:06.022563 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fmz46" Feb 22 00:10:06 crc kubenswrapper[4857]: I0222 00:10:06.682909 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xd4g7" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" containerName="registry-server" probeResult="failure" output=< Feb 22 00:10:06 crc kubenswrapper[4857]: timeout: failed to connect service ":50051" within 1s Feb 22 00:10:06 crc kubenswrapper[4857]: > Feb 22 00:10:07 crc kubenswrapper[4857]: I0222 00:10:07.060759 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fmz46" podUID="299c4a35-6470-4a40-9b90-ad0390497677" containerName="registry-server" probeResult="failure" output=< Feb 22 00:10:07 crc kubenswrapper[4857]: timeout: failed to connect service ":50051" within 1s Feb 22 00:10:07 crc kubenswrapper[4857]: > Feb 22 00:10:07 crc kubenswrapper[4857]: I0222 00:10:07.591089 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" podUID="b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" containerName="oauth-openshift" containerID="cri-o://781a50a624c5adcbde873dbe842de412ec8284a660bd60e0a29009b8a5282985" gracePeriod=15 Feb 22 00:10:08 crc kubenswrapper[4857]: I0222 00:10:08.896328 4857 generic.go:334] "Generic (PLEG): container finished" podID="b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" containerID="781a50a624c5adcbde873dbe842de412ec8284a660bd60e0a29009b8a5282985" exitCode=0 Feb 22 00:10:08 crc kubenswrapper[4857]: I0222 00:10:08.896373 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" event={"ID":"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c","Type":"ContainerDied","Data":"781a50a624c5adcbde873dbe842de412ec8284a660bd60e0a29009b8a5282985"} Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.178669 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.197382 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-idp-0-file-data\") pod \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.197629 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-audit-dir\") pod \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.197849 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sj6pj\" (UniqueName: \"kubernetes.io/projected/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-kube-api-access-sj6pj\") pod \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.197964 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-audit-policies\") pod \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.198268 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-session\") pod \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.198506 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-router-certs\") pod \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.199806 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-template-provider-selection\") pod \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.199928 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-template-login\") pod \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.200002 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-cliconfig\") pod \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.197795 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" (UID: "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.201196 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" (UID: "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.202993 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-ocp-branding-template\") pod \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.203163 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-serving-cert\") pod \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.202526 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" (UID: "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.203352 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-service-ca\") pod \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.203887 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.204017 4857 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.204153 4857 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.203900 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" (UID: "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.207044 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-kube-api-access-sj6pj" (OuterVolumeSpecName: "kube-api-access-sj6pj") pod "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" (UID: "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c"). InnerVolumeSpecName "kube-api-access-sj6pj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.207170 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" (UID: "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.207546 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" (UID: "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.207902 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" (UID: "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.208091 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" (UID: "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.208433 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" (UID: "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.210602 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" (UID: "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.215919 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" (UID: "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.305435 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-template-error\") pod \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.305510 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-trusted-ca-bundle\") pod \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\" (UID: \"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c\") " Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.306141 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.306180 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.306192 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.306202 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.306211 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.306220 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.306233 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sj6pj\" (UniqueName: \"kubernetes.io/projected/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-kube-api-access-sj6pj\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.306244 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.306253 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.306402 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" (UID: "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.319483 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" (UID: "b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.406998 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.407322 4857 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.903078 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" event={"ID":"b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c","Type":"ContainerDied","Data":"42a3dcf05f73f3c5133646209b8c96deec29c498ca560bf228dcf020f53d0cfe"} Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.903123 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lm22m" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.903133 4857 scope.go:117] "RemoveContainer" containerID="781a50a624c5adcbde873dbe842de412ec8284a660bd60e0a29009b8a5282985" Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.931245 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lm22m"] Feb 22 00:10:09 crc kubenswrapper[4857]: I0222 00:10:09.934626 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lm22m"] Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.428544 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.428612 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.428655 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.429325 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7"} pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.429494 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" containerID="cri-o://bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7" gracePeriod=600 Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.673577 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch"] Feb 22 00:10:10 crc kubenswrapper[4857]: E0222 00:10:10.674163 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" containerName="oauth-openshift" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.674177 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" containerName="oauth-openshift" Feb 22 00:10:10 crc kubenswrapper[4857]: E0222 00:10:10.674191 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ced68fc-a805-4595-8043-ce7ec0b802a3" containerName="registry-server" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.674197 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ced68fc-a805-4595-8043-ce7ec0b802a3" containerName="registry-server" Feb 22 00:10:10 crc kubenswrapper[4857]: E0222 00:10:10.674208 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ced68fc-a805-4595-8043-ce7ec0b802a3" containerName="extract-utilities" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.674214 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ced68fc-a805-4595-8043-ce7ec0b802a3" containerName="extract-utilities" Feb 22 00:10:10 crc kubenswrapper[4857]: E0222 00:10:10.674222 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d905b043-15bc-4147-8745-d7c26f087559" containerName="extract-utilities" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.674227 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="d905b043-15bc-4147-8745-d7c26f087559" containerName="extract-utilities" Feb 22 00:10:10 crc kubenswrapper[4857]: E0222 00:10:10.674234 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d905b043-15bc-4147-8745-d7c26f087559" containerName="registry-server" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.674241 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="d905b043-15bc-4147-8745-d7c26f087559" containerName="registry-server" Feb 22 00:10:10 crc kubenswrapper[4857]: E0222 00:10:10.674249 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d905b043-15bc-4147-8745-d7c26f087559" containerName="extract-content" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.674254 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="d905b043-15bc-4147-8745-d7c26f087559" containerName="extract-content" Feb 22 00:10:10 crc kubenswrapper[4857]: E0222 00:10:10.674263 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ced68fc-a805-4595-8043-ce7ec0b802a3" containerName="extract-content" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.674269 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ced68fc-a805-4595-8043-ce7ec0b802a3" containerName="extract-content" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.674378 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" containerName="oauth-openshift" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.674395 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="d905b043-15bc-4147-8745-d7c26f087559" containerName="registry-server" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.674413 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ced68fc-a805-4595-8043-ce7ec0b802a3" containerName="registry-server" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.674799 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.683561 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.683600 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.683996 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.684333 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.684515 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.684561 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.684562 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.684572 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.684580 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.684646 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.684791 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.684936 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.689224 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch"] Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.691119 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.691874 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.696319 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.722295 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-user-template-login\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.722362 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6663279d-6db0-4ee1-81df-13eb906c0d4f-audit-dir\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.722389 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-session\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.722418 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-router-certs\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.722447 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.722472 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.722492 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6663279d-6db0-4ee1-81df-13eb906c0d4f-audit-policies\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.722514 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-user-template-error\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.722535 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq2rv\" (UniqueName: \"kubernetes.io/projected/6663279d-6db0-4ee1-81df-13eb906c0d4f-kube-api-access-tq2rv\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.722836 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.722875 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.722923 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-service-ca\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.722966 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.723002 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.824540 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-router-certs\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.824590 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.824616 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.824644 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6663279d-6db0-4ee1-81df-13eb906c0d4f-audit-policies\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.824674 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-user-template-error\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.824690 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq2rv\" (UniqueName: \"kubernetes.io/projected/6663279d-6db0-4ee1-81df-13eb906c0d4f-kube-api-access-tq2rv\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.824715 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.824731 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.824746 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-service-ca\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.824789 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.824812 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.824836 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-user-template-login\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.824861 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6663279d-6db0-4ee1-81df-13eb906c0d4f-audit-dir\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.824878 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-session\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.825816 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6663279d-6db0-4ee1-81df-13eb906c0d4f-audit-dir\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.826375 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6663279d-6db0-4ee1-81df-13eb906c0d4f-audit-policies\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.826494 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.826773 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.827140 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-service-ca\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.828694 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.829137 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-user-template-login\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.829286 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.829470 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.830360 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.835010 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-user-template-error\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.839349 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-session\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.841308 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6663279d-6db0-4ee1-81df-13eb906c0d4f-v4-0-config-system-router-certs\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.841798 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq2rv\" (UniqueName: \"kubernetes.io/projected/6663279d-6db0-4ee1-81df-13eb906c0d4f-kube-api-access-tq2rv\") pod \"oauth-openshift-5bf578b4ff-ck9ch\" (UID: \"6663279d-6db0-4ee1-81df-13eb906c0d4f\") " pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.911344 4857 generic.go:334] "Generic (PLEG): container finished" podID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerID="bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7" exitCode=0 Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.911417 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" event={"ID":"2ed2fde8-1449-491b-ae21-2fe11ab008ae","Type":"ContainerDied","Data":"bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7"} Feb 22 00:10:10 crc kubenswrapper[4857]: I0222 00:10:10.911447 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" event={"ID":"2ed2fde8-1449-491b-ae21-2fe11ab008ae","Type":"ContainerStarted","Data":"93a06a7512e609878c228069a0a49f796632485175ab77273f65221f6e526dd8"} Feb 22 00:10:11 crc kubenswrapper[4857]: I0222 00:10:11.013085 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:11 crc kubenswrapper[4857]: I0222 00:10:11.083741 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c" path="/var/lib/kubelet/pods/b0ea7458-dfdb-4006-b9bd-2d7e781fdd8c/volumes" Feb 22 00:10:11 crc kubenswrapper[4857]: I0222 00:10:11.175663 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch"] Feb 22 00:10:11 crc kubenswrapper[4857]: W0222 00:10:11.178483 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6663279d_6db0_4ee1_81df_13eb906c0d4f.slice/crio-54642902423fdc95885ceeff818f62a1e1a31ea3827130cde13b16638a64fb2e WatchSource:0}: Error finding container 54642902423fdc95885ceeff818f62a1e1a31ea3827130cde13b16638a64fb2e: Status 404 returned error can't find the container with id 54642902423fdc95885ceeff818f62a1e1a31ea3827130cde13b16638a64fb2e Feb 22 00:10:11 crc kubenswrapper[4857]: I0222 00:10:11.919044 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" event={"ID":"6663279d-6db0-4ee1-81df-13eb906c0d4f","Type":"ContainerStarted","Data":"b6ba83570f9c0f981d0ec139cb97a1df287f8a864a485b9bd61b4de30b0ef9c1"} Feb 22 00:10:11 crc kubenswrapper[4857]: I0222 00:10:11.919409 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" event={"ID":"6663279d-6db0-4ee1-81df-13eb906c0d4f","Type":"ContainerStarted","Data":"54642902423fdc95885ceeff818f62a1e1a31ea3827130cde13b16638a64fb2e"} Feb 22 00:10:11 crc kubenswrapper[4857]: I0222 00:10:11.920609 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:11 crc kubenswrapper[4857]: I0222 00:10:11.936371 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" Feb 22 00:10:11 crc kubenswrapper[4857]: I0222 00:10:11.945371 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5bf578b4ff-ck9ch" podStartSLOduration=29.945355094 podStartE2EDuration="29.945355094s" podCreationTimestamp="2026-02-22 00:09:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:10:11.944183644 +0000 UTC m=+219.582912907" watchObservedRunningTime="2026-02-22 00:10:11.945355094 +0000 UTC m=+219.584084347" Feb 22 00:10:15 crc kubenswrapper[4857]: I0222 00:10:15.041076 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bdmrq" Feb 22 00:10:15 crc kubenswrapper[4857]: I0222 00:10:15.688484 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xd4g7" Feb 22 00:10:15 crc kubenswrapper[4857]: I0222 00:10:15.736616 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xd4g7" Feb 22 00:10:15 crc kubenswrapper[4857]: I0222 00:10:15.907436 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bdmrq"] Feb 22 00:10:15 crc kubenswrapper[4857]: I0222 00:10:15.941128 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bdmrq" podUID="0c14592b-a2d9-46c6-b818-6da2942cb895" containerName="registry-server" containerID="cri-o://f07bd44f1e70a4da4e3ad7e97db6f0ba6754e089402bcc5d506319adeed9d213" gracePeriod=2 Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.063064 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fmz46" Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.105843 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fmz46" Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.371557 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bdmrq" Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.400955 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c14592b-a2d9-46c6-b818-6da2942cb895-catalog-content\") pod \"0c14592b-a2d9-46c6-b818-6da2942cb895\" (UID: \"0c14592b-a2d9-46c6-b818-6da2942cb895\") " Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.401002 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ml52g\" (UniqueName: \"kubernetes.io/projected/0c14592b-a2d9-46c6-b818-6da2942cb895-kube-api-access-ml52g\") pod \"0c14592b-a2d9-46c6-b818-6da2942cb895\" (UID: \"0c14592b-a2d9-46c6-b818-6da2942cb895\") " Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.401050 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c14592b-a2d9-46c6-b818-6da2942cb895-utilities\") pod \"0c14592b-a2d9-46c6-b818-6da2942cb895\" (UID: \"0c14592b-a2d9-46c6-b818-6da2942cb895\") " Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.402176 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c14592b-a2d9-46c6-b818-6da2942cb895-utilities" (OuterVolumeSpecName: "utilities") pod "0c14592b-a2d9-46c6-b818-6da2942cb895" (UID: "0c14592b-a2d9-46c6-b818-6da2942cb895"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.407298 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c14592b-a2d9-46c6-b818-6da2942cb895-kube-api-access-ml52g" (OuterVolumeSpecName: "kube-api-access-ml52g") pod "0c14592b-a2d9-46c6-b818-6da2942cb895" (UID: "0c14592b-a2d9-46c6-b818-6da2942cb895"). InnerVolumeSpecName "kube-api-access-ml52g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.437990 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c14592b-a2d9-46c6-b818-6da2942cb895-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c14592b-a2d9-46c6-b818-6da2942cb895" (UID: "0c14592b-a2d9-46c6-b818-6da2942cb895"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.502927 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c14592b-a2d9-46c6-b818-6da2942cb895-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.503497 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ml52g\" (UniqueName: \"kubernetes.io/projected/0c14592b-a2d9-46c6-b818-6da2942cb895-kube-api-access-ml52g\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.503511 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c14592b-a2d9-46c6-b818-6da2942cb895-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.950026 4857 generic.go:334] "Generic (PLEG): container finished" podID="0c14592b-a2d9-46c6-b818-6da2942cb895" containerID="f07bd44f1e70a4da4e3ad7e97db6f0ba6754e089402bcc5d506319adeed9d213" exitCode=0 Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.950085 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bdmrq" event={"ID":"0c14592b-a2d9-46c6-b818-6da2942cb895","Type":"ContainerDied","Data":"f07bd44f1e70a4da4e3ad7e97db6f0ba6754e089402bcc5d506319adeed9d213"} Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.950137 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bdmrq" event={"ID":"0c14592b-a2d9-46c6-b818-6da2942cb895","Type":"ContainerDied","Data":"9ea747affc4bee529a5ed87989e12c5fd6a9b2f6c28e90b35077f68c3825e094"} Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.950157 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bdmrq" Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.950166 4857 scope.go:117] "RemoveContainer" containerID="f07bd44f1e70a4da4e3ad7e97db6f0ba6754e089402bcc5d506319adeed9d213" Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.968489 4857 scope.go:117] "RemoveContainer" containerID="c5065ca9d55267625ff6c51cfb84762429ca0f4be07e0ba01e2c413ba15f315b" Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.982629 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bdmrq"] Feb 22 00:10:16 crc kubenswrapper[4857]: I0222 00:10:16.986140 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bdmrq"] Feb 22 00:10:17 crc kubenswrapper[4857]: I0222 00:10:17.001718 4857 scope.go:117] "RemoveContainer" containerID="aa9259c83a2418aa714903102d753ff124b035ff728274241cc33a61175681a1" Feb 22 00:10:17 crc kubenswrapper[4857]: I0222 00:10:17.016599 4857 scope.go:117] "RemoveContainer" containerID="f07bd44f1e70a4da4e3ad7e97db6f0ba6754e089402bcc5d506319adeed9d213" Feb 22 00:10:17 crc kubenswrapper[4857]: E0222 00:10:17.017108 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f07bd44f1e70a4da4e3ad7e97db6f0ba6754e089402bcc5d506319adeed9d213\": container with ID starting with f07bd44f1e70a4da4e3ad7e97db6f0ba6754e089402bcc5d506319adeed9d213 not found: ID does not exist" containerID="f07bd44f1e70a4da4e3ad7e97db6f0ba6754e089402bcc5d506319adeed9d213" Feb 22 00:10:17 crc kubenswrapper[4857]: I0222 00:10:17.017158 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f07bd44f1e70a4da4e3ad7e97db6f0ba6754e089402bcc5d506319adeed9d213"} err="failed to get container status \"f07bd44f1e70a4da4e3ad7e97db6f0ba6754e089402bcc5d506319adeed9d213\": rpc error: code = NotFound desc = could not find container \"f07bd44f1e70a4da4e3ad7e97db6f0ba6754e089402bcc5d506319adeed9d213\": container with ID starting with f07bd44f1e70a4da4e3ad7e97db6f0ba6754e089402bcc5d506319adeed9d213 not found: ID does not exist" Feb 22 00:10:17 crc kubenswrapper[4857]: I0222 00:10:17.017185 4857 scope.go:117] "RemoveContainer" containerID="c5065ca9d55267625ff6c51cfb84762429ca0f4be07e0ba01e2c413ba15f315b" Feb 22 00:10:17 crc kubenswrapper[4857]: E0222 00:10:17.017489 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5065ca9d55267625ff6c51cfb84762429ca0f4be07e0ba01e2c413ba15f315b\": container with ID starting with c5065ca9d55267625ff6c51cfb84762429ca0f4be07e0ba01e2c413ba15f315b not found: ID does not exist" containerID="c5065ca9d55267625ff6c51cfb84762429ca0f4be07e0ba01e2c413ba15f315b" Feb 22 00:10:17 crc kubenswrapper[4857]: I0222 00:10:17.017515 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5065ca9d55267625ff6c51cfb84762429ca0f4be07e0ba01e2c413ba15f315b"} err="failed to get container status \"c5065ca9d55267625ff6c51cfb84762429ca0f4be07e0ba01e2c413ba15f315b\": rpc error: code = NotFound desc = could not find container \"c5065ca9d55267625ff6c51cfb84762429ca0f4be07e0ba01e2c413ba15f315b\": container with ID starting with c5065ca9d55267625ff6c51cfb84762429ca0f4be07e0ba01e2c413ba15f315b not found: ID does not exist" Feb 22 00:10:17 crc kubenswrapper[4857]: I0222 00:10:17.017530 4857 scope.go:117] "RemoveContainer" containerID="aa9259c83a2418aa714903102d753ff124b035ff728274241cc33a61175681a1" Feb 22 00:10:17 crc kubenswrapper[4857]: E0222 00:10:17.017869 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa9259c83a2418aa714903102d753ff124b035ff728274241cc33a61175681a1\": container with ID starting with aa9259c83a2418aa714903102d753ff124b035ff728274241cc33a61175681a1 not found: ID does not exist" containerID="aa9259c83a2418aa714903102d753ff124b035ff728274241cc33a61175681a1" Feb 22 00:10:17 crc kubenswrapper[4857]: I0222 00:10:17.017898 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa9259c83a2418aa714903102d753ff124b035ff728274241cc33a61175681a1"} err="failed to get container status \"aa9259c83a2418aa714903102d753ff124b035ff728274241cc33a61175681a1\": rpc error: code = NotFound desc = could not find container \"aa9259c83a2418aa714903102d753ff124b035ff728274241cc33a61175681a1\": container with ID starting with aa9259c83a2418aa714903102d753ff124b035ff728274241cc33a61175681a1 not found: ID does not exist" Feb 22 00:10:17 crc kubenswrapper[4857]: I0222 00:10:17.086900 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c14592b-a2d9-46c6-b818-6da2942cb895" path="/var/lib/kubelet/pods/0c14592b-a2d9-46c6-b818-6da2942cb895/volumes" Feb 22 00:10:18 crc kubenswrapper[4857]: I0222 00:10:18.106478 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fmz46"] Feb 22 00:10:18 crc kubenswrapper[4857]: I0222 00:10:18.106679 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fmz46" podUID="299c4a35-6470-4a40-9b90-ad0390497677" containerName="registry-server" containerID="cri-o://7b5f75d1abc51d34be742c27a04c0b82bcf54997be03024b6f687faf71696113" gracePeriod=2 Feb 22 00:10:18 crc kubenswrapper[4857]: I0222 00:10:18.438894 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fmz46" Feb 22 00:10:18 crc kubenswrapper[4857]: I0222 00:10:18.524281 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/299c4a35-6470-4a40-9b90-ad0390497677-catalog-content\") pod \"299c4a35-6470-4a40-9b90-ad0390497677\" (UID: \"299c4a35-6470-4a40-9b90-ad0390497677\") " Feb 22 00:10:18 crc kubenswrapper[4857]: I0222 00:10:18.524348 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9fqw\" (UniqueName: \"kubernetes.io/projected/299c4a35-6470-4a40-9b90-ad0390497677-kube-api-access-k9fqw\") pod \"299c4a35-6470-4a40-9b90-ad0390497677\" (UID: \"299c4a35-6470-4a40-9b90-ad0390497677\") " Feb 22 00:10:18 crc kubenswrapper[4857]: I0222 00:10:18.524392 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/299c4a35-6470-4a40-9b90-ad0390497677-utilities\") pod \"299c4a35-6470-4a40-9b90-ad0390497677\" (UID: \"299c4a35-6470-4a40-9b90-ad0390497677\") " Feb 22 00:10:18 crc kubenswrapper[4857]: I0222 00:10:18.525456 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/299c4a35-6470-4a40-9b90-ad0390497677-utilities" (OuterVolumeSpecName: "utilities") pod "299c4a35-6470-4a40-9b90-ad0390497677" (UID: "299c4a35-6470-4a40-9b90-ad0390497677"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:10:18 crc kubenswrapper[4857]: I0222 00:10:18.530276 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/299c4a35-6470-4a40-9b90-ad0390497677-kube-api-access-k9fqw" (OuterVolumeSpecName: "kube-api-access-k9fqw") pod "299c4a35-6470-4a40-9b90-ad0390497677" (UID: "299c4a35-6470-4a40-9b90-ad0390497677"). InnerVolumeSpecName "kube-api-access-k9fqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:10:18 crc kubenswrapper[4857]: I0222 00:10:18.625926 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/299c4a35-6470-4a40-9b90-ad0390497677-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:18 crc kubenswrapper[4857]: I0222 00:10:18.625982 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9fqw\" (UniqueName: \"kubernetes.io/projected/299c4a35-6470-4a40-9b90-ad0390497677-kube-api-access-k9fqw\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:18 crc kubenswrapper[4857]: I0222 00:10:18.637334 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/299c4a35-6470-4a40-9b90-ad0390497677-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "299c4a35-6470-4a40-9b90-ad0390497677" (UID: "299c4a35-6470-4a40-9b90-ad0390497677"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:10:18 crc kubenswrapper[4857]: I0222 00:10:18.727562 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/299c4a35-6470-4a40-9b90-ad0390497677-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:18 crc kubenswrapper[4857]: I0222 00:10:18.967907 4857 generic.go:334] "Generic (PLEG): container finished" podID="299c4a35-6470-4a40-9b90-ad0390497677" containerID="7b5f75d1abc51d34be742c27a04c0b82bcf54997be03024b6f687faf71696113" exitCode=0 Feb 22 00:10:18 crc kubenswrapper[4857]: I0222 00:10:18.967994 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmz46" event={"ID":"299c4a35-6470-4a40-9b90-ad0390497677","Type":"ContainerDied","Data":"7b5f75d1abc51d34be742c27a04c0b82bcf54997be03024b6f687faf71696113"} Feb 22 00:10:18 crc kubenswrapper[4857]: I0222 00:10:18.968078 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fmz46" event={"ID":"299c4a35-6470-4a40-9b90-ad0390497677","Type":"ContainerDied","Data":"1f4d049e289f91deafbff62b2081acf0666bd799d7f02c2633ea706901633011"} Feb 22 00:10:18 crc kubenswrapper[4857]: I0222 00:10:18.968119 4857 scope.go:117] "RemoveContainer" containerID="7b5f75d1abc51d34be742c27a04c0b82bcf54997be03024b6f687faf71696113" Feb 22 00:10:18 crc kubenswrapper[4857]: I0222 00:10:18.968318 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fmz46" Feb 22 00:10:18 crc kubenswrapper[4857]: I0222 00:10:18.987546 4857 scope.go:117] "RemoveContainer" containerID="971edb974894b8c9cbd48f55ca1d5493160884bc60c723186931b064bae04712" Feb 22 00:10:19 crc kubenswrapper[4857]: I0222 00:10:19.005069 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fmz46"] Feb 22 00:10:19 crc kubenswrapper[4857]: I0222 00:10:19.006915 4857 scope.go:117] "RemoveContainer" containerID="5fe027cb577b65b59ae8a4d1e15b4d33b7276edd73f93df2e2a9de478d1cf8cb" Feb 22 00:10:19 crc kubenswrapper[4857]: I0222 00:10:19.007639 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fmz46"] Feb 22 00:10:19 crc kubenswrapper[4857]: I0222 00:10:19.035756 4857 scope.go:117] "RemoveContainer" containerID="7b5f75d1abc51d34be742c27a04c0b82bcf54997be03024b6f687faf71696113" Feb 22 00:10:19 crc kubenswrapper[4857]: E0222 00:10:19.036127 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b5f75d1abc51d34be742c27a04c0b82bcf54997be03024b6f687faf71696113\": container with ID starting with 7b5f75d1abc51d34be742c27a04c0b82bcf54997be03024b6f687faf71696113 not found: ID does not exist" containerID="7b5f75d1abc51d34be742c27a04c0b82bcf54997be03024b6f687faf71696113" Feb 22 00:10:19 crc kubenswrapper[4857]: I0222 00:10:19.036198 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b5f75d1abc51d34be742c27a04c0b82bcf54997be03024b6f687faf71696113"} err="failed to get container status \"7b5f75d1abc51d34be742c27a04c0b82bcf54997be03024b6f687faf71696113\": rpc error: code = NotFound desc = could not find container \"7b5f75d1abc51d34be742c27a04c0b82bcf54997be03024b6f687faf71696113\": container with ID starting with 7b5f75d1abc51d34be742c27a04c0b82bcf54997be03024b6f687faf71696113 not found: ID does not exist" Feb 22 00:10:19 crc kubenswrapper[4857]: I0222 00:10:19.036238 4857 scope.go:117] "RemoveContainer" containerID="971edb974894b8c9cbd48f55ca1d5493160884bc60c723186931b064bae04712" Feb 22 00:10:19 crc kubenswrapper[4857]: E0222 00:10:19.036584 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"971edb974894b8c9cbd48f55ca1d5493160884bc60c723186931b064bae04712\": container with ID starting with 971edb974894b8c9cbd48f55ca1d5493160884bc60c723186931b064bae04712 not found: ID does not exist" containerID="971edb974894b8c9cbd48f55ca1d5493160884bc60c723186931b064bae04712" Feb 22 00:10:19 crc kubenswrapper[4857]: I0222 00:10:19.036616 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"971edb974894b8c9cbd48f55ca1d5493160884bc60c723186931b064bae04712"} err="failed to get container status \"971edb974894b8c9cbd48f55ca1d5493160884bc60c723186931b064bae04712\": rpc error: code = NotFound desc = could not find container \"971edb974894b8c9cbd48f55ca1d5493160884bc60c723186931b064bae04712\": container with ID starting with 971edb974894b8c9cbd48f55ca1d5493160884bc60c723186931b064bae04712 not found: ID does not exist" Feb 22 00:10:19 crc kubenswrapper[4857]: I0222 00:10:19.036638 4857 scope.go:117] "RemoveContainer" containerID="5fe027cb577b65b59ae8a4d1e15b4d33b7276edd73f93df2e2a9de478d1cf8cb" Feb 22 00:10:19 crc kubenswrapper[4857]: E0222 00:10:19.036818 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fe027cb577b65b59ae8a4d1e15b4d33b7276edd73f93df2e2a9de478d1cf8cb\": container with ID starting with 5fe027cb577b65b59ae8a4d1e15b4d33b7276edd73f93df2e2a9de478d1cf8cb not found: ID does not exist" containerID="5fe027cb577b65b59ae8a4d1e15b4d33b7276edd73f93df2e2a9de478d1cf8cb" Feb 22 00:10:19 crc kubenswrapper[4857]: I0222 00:10:19.036841 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fe027cb577b65b59ae8a4d1e15b4d33b7276edd73f93df2e2a9de478d1cf8cb"} err="failed to get container status \"5fe027cb577b65b59ae8a4d1e15b4d33b7276edd73f93df2e2a9de478d1cf8cb\": rpc error: code = NotFound desc = could not find container \"5fe027cb577b65b59ae8a4d1e15b4d33b7276edd73f93df2e2a9de478d1cf8cb\": container with ID starting with 5fe027cb577b65b59ae8a4d1e15b4d33b7276edd73f93df2e2a9de478d1cf8cb not found: ID does not exist" Feb 22 00:10:19 crc kubenswrapper[4857]: I0222 00:10:19.085469 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="299c4a35-6470-4a40-9b90-ad0390497677" path="/var/lib/kubelet/pods/299c4a35-6470-4a40-9b90-ad0390497677/volumes" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.553408 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m5rft"] Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.555315 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m5rft" podUID="18611d84-22b3-49df-9569-ac383eb42cab" containerName="registry-server" containerID="cri-o://7ba7afef408b9c9fe34536b3e36156921afe7f08cf8f984219ea7dc53fbd01cf" gracePeriod=30 Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.561018 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5rpdr"] Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.561323 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5rpdr" podUID="fa45663f-d64c-431c-be1d-29175cc804ea" containerName="registry-server" containerID="cri-o://0d3f598c3d03bf8ce8f4bef7c23fdabe3795b3736cc7a62fc550d7a372dca74c" gracePeriod=30 Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.569803 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vmk2g"] Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.569969 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" podUID="4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" containerName="marketplace-operator" containerID="cri-o://1837b66f71bf7805f8b4307a67a79a46079d745062c0638df7ba39b0091df73b" gracePeriod=30 Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.577992 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-797dx"] Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.578229 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-797dx" podUID="10f2d337-12f1-4949-9872-08df9aa084ab" containerName="registry-server" containerID="cri-o://5d7a76f88ed88af6871dfb7951a544360508850bbf76aab186a86107c836c6fb" gracePeriod=30 Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.589586 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xd4g7"] Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.589645 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4cnjs"] Feb 22 00:10:29 crc kubenswrapper[4857]: E0222 00:10:29.589819 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c14592b-a2d9-46c6-b818-6da2942cb895" containerName="extract-utilities" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.589830 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c14592b-a2d9-46c6-b818-6da2942cb895" containerName="extract-utilities" Feb 22 00:10:29 crc kubenswrapper[4857]: E0222 00:10:29.589843 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c14592b-a2d9-46c6-b818-6da2942cb895" containerName="registry-server" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.589849 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c14592b-a2d9-46c6-b818-6da2942cb895" containerName="registry-server" Feb 22 00:10:29 crc kubenswrapper[4857]: E0222 00:10:29.589859 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="299c4a35-6470-4a40-9b90-ad0390497677" containerName="extract-content" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.589865 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="299c4a35-6470-4a40-9b90-ad0390497677" containerName="extract-content" Feb 22 00:10:29 crc kubenswrapper[4857]: E0222 00:10:29.589871 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="299c4a35-6470-4a40-9b90-ad0390497677" containerName="registry-server" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.589876 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="299c4a35-6470-4a40-9b90-ad0390497677" containerName="registry-server" Feb 22 00:10:29 crc kubenswrapper[4857]: E0222 00:10:29.589891 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="299c4a35-6470-4a40-9b90-ad0390497677" containerName="extract-utilities" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.589896 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="299c4a35-6470-4a40-9b90-ad0390497677" containerName="extract-utilities" Feb 22 00:10:29 crc kubenswrapper[4857]: E0222 00:10:29.589906 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c14592b-a2d9-46c6-b818-6da2942cb895" containerName="extract-content" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.589912 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c14592b-a2d9-46c6-b818-6da2942cb895" containerName="extract-content" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.589990 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c14592b-a2d9-46c6-b818-6da2942cb895" containerName="registry-server" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.590003 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="299c4a35-6470-4a40-9b90-ad0390497677" containerName="registry-server" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.590196 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xd4g7" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" containerName="registry-server" containerID="cri-o://34c2234b5cda4e087d988f3152f4a10fc71a296bec05bfffd187d84e66ce1cf5" gracePeriod=30 Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.594331 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.602473 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4cnjs"] Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.665132 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aec64a2e-ff83-4d35-a6cc-691c906ee0d9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4cnjs\" (UID: \"aec64a2e-ff83-4d35-a6cc-691c906ee0d9\") " pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.665205 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stzg9\" (UniqueName: \"kubernetes.io/projected/aec64a2e-ff83-4d35-a6cc-691c906ee0d9-kube-api-access-stzg9\") pod \"marketplace-operator-79b997595-4cnjs\" (UID: \"aec64a2e-ff83-4d35-a6cc-691c906ee0d9\") " pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.665223 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/aec64a2e-ff83-4d35-a6cc-691c906ee0d9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4cnjs\" (UID: \"aec64a2e-ff83-4d35-a6cc-691c906ee0d9\") " pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.765888 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aec64a2e-ff83-4d35-a6cc-691c906ee0d9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4cnjs\" (UID: \"aec64a2e-ff83-4d35-a6cc-691c906ee0d9\") " pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.766305 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stzg9\" (UniqueName: \"kubernetes.io/projected/aec64a2e-ff83-4d35-a6cc-691c906ee0d9-kube-api-access-stzg9\") pod \"marketplace-operator-79b997595-4cnjs\" (UID: \"aec64a2e-ff83-4d35-a6cc-691c906ee0d9\") " pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.766332 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/aec64a2e-ff83-4d35-a6cc-691c906ee0d9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4cnjs\" (UID: \"aec64a2e-ff83-4d35-a6cc-691c906ee0d9\") " pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.767568 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/aec64a2e-ff83-4d35-a6cc-691c906ee0d9-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4cnjs\" (UID: \"aec64a2e-ff83-4d35-a6cc-691c906ee0d9\") " pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.785209 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/aec64a2e-ff83-4d35-a6cc-691c906ee0d9-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4cnjs\" (UID: \"aec64a2e-ff83-4d35-a6cc-691c906ee0d9\") " pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.786600 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stzg9\" (UniqueName: \"kubernetes.io/projected/aec64a2e-ff83-4d35-a6cc-691c906ee0d9-kube-api-access-stzg9\") pod \"marketplace-operator-79b997595-4cnjs\" (UID: \"aec64a2e-ff83-4d35-a6cc-691c906ee0d9\") " pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.983054 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.988768 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rpdr" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.990389 4857 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.990678 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df" gracePeriod=15 Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.990811 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918" gracePeriod=15 Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.990873 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa" gracePeriod=15 Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.990918 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f" gracePeriod=15 Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.991270 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45" gracePeriod=15 Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.993839 4857 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 22 00:10:29 crc kubenswrapper[4857]: E0222 00:10:29.994021 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994057 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 22 00:10:29 crc kubenswrapper[4857]: E0222 00:10:29.994070 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994077 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 22 00:10:29 crc kubenswrapper[4857]: E0222 00:10:29.994085 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994090 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 22 00:10:29 crc kubenswrapper[4857]: E0222 00:10:29.994097 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994103 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 22 00:10:29 crc kubenswrapper[4857]: E0222 00:10:29.994111 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994118 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 22 00:10:29 crc kubenswrapper[4857]: E0222 00:10:29.994127 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994133 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 22 00:10:29 crc kubenswrapper[4857]: E0222 00:10:29.994143 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa45663f-d64c-431c-be1d-29175cc804ea" containerName="extract-content" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994148 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa45663f-d64c-431c-be1d-29175cc804ea" containerName="extract-content" Feb 22 00:10:29 crc kubenswrapper[4857]: E0222 00:10:29.994157 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994163 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 22 00:10:29 crc kubenswrapper[4857]: E0222 00:10:29.994175 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa45663f-d64c-431c-be1d-29175cc804ea" containerName="extract-utilities" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994183 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa45663f-d64c-431c-be1d-29175cc804ea" containerName="extract-utilities" Feb 22 00:10:29 crc kubenswrapper[4857]: E0222 00:10:29.994192 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa45663f-d64c-431c-be1d-29175cc804ea" containerName="registry-server" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994200 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa45663f-d64c-431c-be1d-29175cc804ea" containerName="registry-server" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994283 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994294 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa45663f-d64c-431c-be1d-29175cc804ea" containerName="registry-server" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994300 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994307 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994315 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994322 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994328 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 22 00:10:29 crc kubenswrapper[4857]: E0222 00:10:29.994420 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994427 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 22 00:10:29 crc kubenswrapper[4857]: I0222 00:10:29.994506 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.005080 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.005411 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5rft" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.016363 4857 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.016571 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" containerName="marketplace-operator" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.016587 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" containerName="marketplace-operator" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.016600 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18611d84-22b3-49df-9569-ac383eb42cab" containerName="extract-content" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.016625 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="18611d84-22b3-49df-9569-ac383eb42cab" containerName="extract-content" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.016634 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18611d84-22b3-49df-9569-ac383eb42cab" containerName="registry-server" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.016639 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="18611d84-22b3-49df-9569-ac383eb42cab" containerName="registry-server" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.016655 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18611d84-22b3-49df-9569-ac383eb42cab" containerName="extract-utilities" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.016660 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="18611d84-22b3-49df-9569-ac383eb42cab" containerName="extract-utilities" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.016740 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" containerName="marketplace-operator" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.016752 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="18611d84-22b3-49df-9569-ac383eb42cab" containerName="registry-server" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.017150 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.023453 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-797dx" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.029131 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xd4g7" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.065263 4857 generic.go:334] "Generic (PLEG): container finished" podID="10f2d337-12f1-4949-9872-08df9aa084ab" containerID="5d7a76f88ed88af6871dfb7951a544360508850bbf76aab186a86107c836c6fb" exitCode=0 Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.065381 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-797dx" event={"ID":"10f2d337-12f1-4949-9872-08df9aa084ab","Type":"ContainerDied","Data":"5d7a76f88ed88af6871dfb7951a544360508850bbf76aab186a86107c836c6fb"} Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.065415 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-797dx" event={"ID":"10f2d337-12f1-4949-9872-08df9aa084ab","Type":"ContainerDied","Data":"702273ed60bd457682a58cb48d335b676624d61f9b8253808e3653e3837389ea"} Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.065447 4857 scope.go:117] "RemoveContainer" containerID="5d7a76f88ed88af6871dfb7951a544360508850bbf76aab186a86107c836c6fb" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.066141 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-797dx" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.081479 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5rft" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.081282 4857 generic.go:334] "Generic (PLEG): container finished" podID="18611d84-22b3-49df-9569-ac383eb42cab" containerID="7ba7afef408b9c9fe34536b3e36156921afe7f08cf8f984219ea7dc53fbd01cf" exitCode=0 Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.081723 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5rft" event={"ID":"18611d84-22b3-49df-9569-ac383eb42cab","Type":"ContainerDied","Data":"7ba7afef408b9c9fe34536b3e36156921afe7f08cf8f984219ea7dc53fbd01cf"} Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.082215 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5rft" event={"ID":"18611d84-22b3-49df-9569-ac383eb42cab","Type":"ContainerDied","Data":"d16007b43d7e06f1119c6c560c28310c5bed58bc9a79575ffb0b2c94354c1c33"} Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.149760 4857 generic.go:334] "Generic (PLEG): container finished" podID="fa45663f-d64c-431c-be1d-29175cc804ea" containerID="0d3f598c3d03bf8ce8f4bef7c23fdabe3795b3736cc7a62fc550d7a372dca74c" exitCode=0 Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.150420 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rpdr" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.150585 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rpdr" event={"ID":"fa45663f-d64c-431c-be1d-29175cc804ea","Type":"ContainerDied","Data":"0d3f598c3d03bf8ce8f4bef7c23fdabe3795b3736cc7a62fc550d7a372dca74c"} Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.150617 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rpdr" event={"ID":"fa45663f-d64c-431c-be1d-29175cc804ea","Type":"ContainerDied","Data":"64daa7086566ca522beeda500051a89d0e4adc6bc1b84155280bd8a2178c446c"} Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.156393 4857 generic.go:334] "Generic (PLEG): container finished" podID="4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" containerID="1837b66f71bf7805f8b4307a67a79a46079d745062c0638df7ba39b0091df73b" exitCode=0 Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.156487 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.156644 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" event={"ID":"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8","Type":"ContainerDied","Data":"1837b66f71bf7805f8b4307a67a79a46079d745062c0638df7ba39b0091df73b"} Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.156688 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" event={"ID":"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8","Type":"ContainerDied","Data":"7c7fd9973de16e4423d230254b9ba870314175413f1a2911b5e17d6da927271e"} Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.164227 4857 generic.go:334] "Generic (PLEG): container finished" podID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" containerID="34c2234b5cda4e087d988f3152f4a10fc71a296bec05bfffd187d84e66ce1cf5" exitCode=0 Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.164261 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xd4g7" event={"ID":"6d73807b-5e0c-4607-8393-5123a6cdbf9b","Type":"ContainerDied","Data":"34c2234b5cda4e087d988f3152f4a10fc71a296bec05bfffd187d84e66ce1cf5"} Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.164307 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xd4g7" event={"ID":"6d73807b-5e0c-4607-8393-5123a6cdbf9b","Type":"ContainerDied","Data":"f53f3c846ec1ce2552bb9b740f4929b3c4292716d057a1378b9f20362d315934"} Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.164394 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xd4g7" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.169631 4857 scope.go:117] "RemoveContainer" containerID="d2baba890384c306149473909fe2096c933348ac570a9a5e78479881563b25f3" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.182856 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhpbm\" (UniqueName: \"kubernetes.io/projected/fa45663f-d64c-431c-be1d-29175cc804ea-kube-api-access-mhpbm\") pod \"fa45663f-d64c-431c-be1d-29175cc804ea\" (UID: \"fa45663f-d64c-431c-be1d-29175cc804ea\") " Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.182897 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8-marketplace-operator-metrics\") pod \"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8\" (UID: \"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8\") " Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.182919 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18611d84-22b3-49df-9569-ac383eb42cab-utilities\") pod \"18611d84-22b3-49df-9569-ac383eb42cab\" (UID: \"18611d84-22b3-49df-9569-ac383eb42cab\") " Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.182956 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cd2qp\" (UniqueName: \"kubernetes.io/projected/6d73807b-5e0c-4607-8393-5123a6cdbf9b-kube-api-access-cd2qp\") pod \"6d73807b-5e0c-4607-8393-5123a6cdbf9b\" (UID: \"6d73807b-5e0c-4607-8393-5123a6cdbf9b\") " Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.182976 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa45663f-d64c-431c-be1d-29175cc804ea-utilities\") pod \"fa45663f-d64c-431c-be1d-29175cc804ea\" (UID: \"fa45663f-d64c-431c-be1d-29175cc804ea\") " Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.182999 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f2d337-12f1-4949-9872-08df9aa084ab-catalog-content\") pod \"10f2d337-12f1-4949-9872-08df9aa084ab\" (UID: \"10f2d337-12f1-4949-9872-08df9aa084ab\") " Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.183021 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59kjl\" (UniqueName: \"kubernetes.io/projected/18611d84-22b3-49df-9569-ac383eb42cab-kube-api-access-59kjl\") pod \"18611d84-22b3-49df-9569-ac383eb42cab\" (UID: \"18611d84-22b3-49df-9569-ac383eb42cab\") " Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.183058 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8-marketplace-trusted-ca\") pod \"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8\" (UID: \"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8\") " Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.183077 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d73807b-5e0c-4607-8393-5123a6cdbf9b-utilities\") pod \"6d73807b-5e0c-4607-8393-5123a6cdbf9b\" (UID: \"6d73807b-5e0c-4607-8393-5123a6cdbf9b\") " Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.183101 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d73807b-5e0c-4607-8393-5123a6cdbf9b-catalog-content\") pod \"6d73807b-5e0c-4607-8393-5123a6cdbf9b\" (UID: \"6d73807b-5e0c-4607-8393-5123a6cdbf9b\") " Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.183116 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f2d337-12f1-4949-9872-08df9aa084ab-utilities\") pod \"10f2d337-12f1-4949-9872-08df9aa084ab\" (UID: \"10f2d337-12f1-4949-9872-08df9aa084ab\") " Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.183143 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8v27\" (UniqueName: \"kubernetes.io/projected/10f2d337-12f1-4949-9872-08df9aa084ab-kube-api-access-f8v27\") pod \"10f2d337-12f1-4949-9872-08df9aa084ab\" (UID: \"10f2d337-12f1-4949-9872-08df9aa084ab\") " Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.183161 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18611d84-22b3-49df-9569-ac383eb42cab-catalog-content\") pod \"18611d84-22b3-49df-9569-ac383eb42cab\" (UID: \"18611d84-22b3-49df-9569-ac383eb42cab\") " Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.183182 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtzzn\" (UniqueName: \"kubernetes.io/projected/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8-kube-api-access-qtzzn\") pod \"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8\" (UID: \"4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8\") " Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.183206 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa45663f-d64c-431c-be1d-29175cc804ea-catalog-content\") pod \"fa45663f-d64c-431c-be1d-29175cc804ea\" (UID: \"fa45663f-d64c-431c-be1d-29175cc804ea\") " Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.183314 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.183344 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.183362 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.183376 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.183399 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.183416 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.183452 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.183466 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.185612 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18611d84-22b3-49df-9569-ac383eb42cab-utilities" (OuterVolumeSpecName: "utilities") pod "18611d84-22b3-49df-9569-ac383eb42cab" (UID: "18611d84-22b3-49df-9569-ac383eb42cab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.188101 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10f2d337-12f1-4949-9872-08df9aa084ab-utilities" (OuterVolumeSpecName: "utilities") pod "10f2d337-12f1-4949-9872-08df9aa084ab" (UID: "10f2d337-12f1-4949-9872-08df9aa084ab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.188971 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d73807b-5e0c-4607-8393-5123a6cdbf9b-utilities" (OuterVolumeSpecName: "utilities") pod "6d73807b-5e0c-4607-8393-5123a6cdbf9b" (UID: "6d73807b-5e0c-4607-8393-5123a6cdbf9b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.189370 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" (UID: "4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.190084 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa45663f-d64c-431c-be1d-29175cc804ea-utilities" (OuterVolumeSpecName: "utilities") pod "fa45663f-d64c-431c-be1d-29175cc804ea" (UID: "fa45663f-d64c-431c-be1d-29175cc804ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.190472 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d73807b-5e0c-4607-8393-5123a6cdbf9b-kube-api-access-cd2qp" (OuterVolumeSpecName: "kube-api-access-cd2qp") pod "6d73807b-5e0c-4607-8393-5123a6cdbf9b" (UID: "6d73807b-5e0c-4607-8393-5123a6cdbf9b"). InnerVolumeSpecName "kube-api-access-cd2qp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.190727 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" (UID: "4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.193615 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8-kube-api-access-qtzzn" (OuterVolumeSpecName: "kube-api-access-qtzzn") pod "4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" (UID: "4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8"). InnerVolumeSpecName "kube-api-access-qtzzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.195297 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10f2d337-12f1-4949-9872-08df9aa084ab-kube-api-access-f8v27" (OuterVolumeSpecName: "kube-api-access-f8v27") pod "10f2d337-12f1-4949-9872-08df9aa084ab" (UID: "10f2d337-12f1-4949-9872-08df9aa084ab"). InnerVolumeSpecName "kube-api-access-f8v27". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.195920 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18611d84-22b3-49df-9569-ac383eb42cab-kube-api-access-59kjl" (OuterVolumeSpecName: "kube-api-access-59kjl") pod "18611d84-22b3-49df-9569-ac383eb42cab" (UID: "18611d84-22b3-49df-9569-ac383eb42cab"). InnerVolumeSpecName "kube-api-access-59kjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.221132 4857 scope.go:117] "RemoveContainer" containerID="c035cf724c76965a182f07533df88b7f93333062ee13f7d63bd2843ba86a7173" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.225337 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa45663f-d64c-431c-be1d-29175cc804ea-kube-api-access-mhpbm" (OuterVolumeSpecName: "kube-api-access-mhpbm") pod "fa45663f-d64c-431c-be1d-29175cc804ea" (UID: "fa45663f-d64c-431c-be1d-29175cc804ea"). InnerVolumeSpecName "kube-api-access-mhpbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.240309 4857 scope.go:117] "RemoveContainer" containerID="5d7a76f88ed88af6871dfb7951a544360508850bbf76aab186a86107c836c6fb" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.240876 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d7a76f88ed88af6871dfb7951a544360508850bbf76aab186a86107c836c6fb\": container with ID starting with 5d7a76f88ed88af6871dfb7951a544360508850bbf76aab186a86107c836c6fb not found: ID does not exist" containerID="5d7a76f88ed88af6871dfb7951a544360508850bbf76aab186a86107c836c6fb" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.240913 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d7a76f88ed88af6871dfb7951a544360508850bbf76aab186a86107c836c6fb"} err="failed to get container status \"5d7a76f88ed88af6871dfb7951a544360508850bbf76aab186a86107c836c6fb\": rpc error: code = NotFound desc = could not find container \"5d7a76f88ed88af6871dfb7951a544360508850bbf76aab186a86107c836c6fb\": container with ID starting with 5d7a76f88ed88af6871dfb7951a544360508850bbf76aab186a86107c836c6fb not found: ID does not exist" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.240942 4857 scope.go:117] "RemoveContainer" containerID="d2baba890384c306149473909fe2096c933348ac570a9a5e78479881563b25f3" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.241366 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2baba890384c306149473909fe2096c933348ac570a9a5e78479881563b25f3\": container with ID starting with d2baba890384c306149473909fe2096c933348ac570a9a5e78479881563b25f3 not found: ID does not exist" containerID="d2baba890384c306149473909fe2096c933348ac570a9a5e78479881563b25f3" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.241403 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2baba890384c306149473909fe2096c933348ac570a9a5e78479881563b25f3"} err="failed to get container status \"d2baba890384c306149473909fe2096c933348ac570a9a5e78479881563b25f3\": rpc error: code = NotFound desc = could not find container \"d2baba890384c306149473909fe2096c933348ac570a9a5e78479881563b25f3\": container with ID starting with d2baba890384c306149473909fe2096c933348ac570a9a5e78479881563b25f3 not found: ID does not exist" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.241431 4857 scope.go:117] "RemoveContainer" containerID="c035cf724c76965a182f07533df88b7f93333062ee13f7d63bd2843ba86a7173" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.241722 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c035cf724c76965a182f07533df88b7f93333062ee13f7d63bd2843ba86a7173\": container with ID starting with c035cf724c76965a182f07533df88b7f93333062ee13f7d63bd2843ba86a7173 not found: ID does not exist" containerID="c035cf724c76965a182f07533df88b7f93333062ee13f7d63bd2843ba86a7173" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.241740 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c035cf724c76965a182f07533df88b7f93333062ee13f7d63bd2843ba86a7173"} err="failed to get container status \"c035cf724c76965a182f07533df88b7f93333062ee13f7d63bd2843ba86a7173\": rpc error: code = NotFound desc = could not find container \"c035cf724c76965a182f07533df88b7f93333062ee13f7d63bd2843ba86a7173\": container with ID starting with c035cf724c76965a182f07533df88b7f93333062ee13f7d63bd2843ba86a7173 not found: ID does not exist" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.241754 4857 scope.go:117] "RemoveContainer" containerID="7ba7afef408b9c9fe34536b3e36156921afe7f08cf8f984219ea7dc53fbd01cf" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.249631 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10f2d337-12f1-4949-9872-08df9aa084ab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10f2d337-12f1-4949-9872-08df9aa084ab" (UID: "10f2d337-12f1-4949-9872-08df9aa084ab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.258077 4857 scope.go:117] "RemoveContainer" containerID="d807fcd19722a12043390537664dbf2f93cdad07a48817b2d3130ba55755dd6d" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.266105 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18611d84-22b3-49df-9569-ac383eb42cab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18611d84-22b3-49df-9569-ac383eb42cab" (UID: "18611d84-22b3-49df-9569-ac383eb42cab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.267496 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa45663f-d64c-431c-be1d-29175cc804ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa45663f-d64c-431c-be1d-29175cc804ea" (UID: "fa45663f-d64c-431c-be1d-29175cc804ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.283282 4857 scope.go:117] "RemoveContainer" containerID="ee5d8876d1488183eb5edf35a9efa3293b59bfcb8d52c5fa2743306c94e399f5" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286608 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286649 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286676 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286697 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286726 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286740 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286785 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286803 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286859 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8v27\" (UniqueName: \"kubernetes.io/projected/10f2d337-12f1-4949-9872-08df9aa084ab-kube-api-access-f8v27\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286873 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18611d84-22b3-49df-9569-ac383eb42cab-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286885 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtzzn\" (UniqueName: \"kubernetes.io/projected/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8-kube-api-access-qtzzn\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286897 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa45663f-d64c-431c-be1d-29175cc804ea-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286908 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhpbm\" (UniqueName: \"kubernetes.io/projected/fa45663f-d64c-431c-be1d-29175cc804ea-kube-api-access-mhpbm\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286916 4857 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286926 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18611d84-22b3-49df-9569-ac383eb42cab-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286936 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cd2qp\" (UniqueName: \"kubernetes.io/projected/6d73807b-5e0c-4607-8393-5123a6cdbf9b-kube-api-access-cd2qp\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286947 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa45663f-d64c-431c-be1d-29175cc804ea-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286960 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10f2d337-12f1-4949-9872-08df9aa084ab-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286971 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59kjl\" (UniqueName: \"kubernetes.io/projected/18611d84-22b3-49df-9569-ac383eb42cab-kube-api-access-59kjl\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286982 4857 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.286993 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d73807b-5e0c-4607-8393-5123a6cdbf9b-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.287003 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10f2d337-12f1-4949-9872-08df9aa084ab-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.287065 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.287104 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.287133 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.287161 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.287186 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.287211 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.287239 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.287266 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.298806 4857 scope.go:117] "RemoveContainer" containerID="7ba7afef408b9c9fe34536b3e36156921afe7f08cf8f984219ea7dc53fbd01cf" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.301446 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ba7afef408b9c9fe34536b3e36156921afe7f08cf8f984219ea7dc53fbd01cf\": container with ID starting with 7ba7afef408b9c9fe34536b3e36156921afe7f08cf8f984219ea7dc53fbd01cf not found: ID does not exist" containerID="7ba7afef408b9c9fe34536b3e36156921afe7f08cf8f984219ea7dc53fbd01cf" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.301500 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ba7afef408b9c9fe34536b3e36156921afe7f08cf8f984219ea7dc53fbd01cf"} err="failed to get container status \"7ba7afef408b9c9fe34536b3e36156921afe7f08cf8f984219ea7dc53fbd01cf\": rpc error: code = NotFound desc = could not find container \"7ba7afef408b9c9fe34536b3e36156921afe7f08cf8f984219ea7dc53fbd01cf\": container with ID starting with 7ba7afef408b9c9fe34536b3e36156921afe7f08cf8f984219ea7dc53fbd01cf not found: ID does not exist" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.301537 4857 scope.go:117] "RemoveContainer" containerID="d807fcd19722a12043390537664dbf2f93cdad07a48817b2d3130ba55755dd6d" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.302001 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d807fcd19722a12043390537664dbf2f93cdad07a48817b2d3130ba55755dd6d\": container with ID starting with d807fcd19722a12043390537664dbf2f93cdad07a48817b2d3130ba55755dd6d not found: ID does not exist" containerID="d807fcd19722a12043390537664dbf2f93cdad07a48817b2d3130ba55755dd6d" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.302052 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d807fcd19722a12043390537664dbf2f93cdad07a48817b2d3130ba55755dd6d"} err="failed to get container status \"d807fcd19722a12043390537664dbf2f93cdad07a48817b2d3130ba55755dd6d\": rpc error: code = NotFound desc = could not find container \"d807fcd19722a12043390537664dbf2f93cdad07a48817b2d3130ba55755dd6d\": container with ID starting with d807fcd19722a12043390537664dbf2f93cdad07a48817b2d3130ba55755dd6d not found: ID does not exist" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.302085 4857 scope.go:117] "RemoveContainer" containerID="ee5d8876d1488183eb5edf35a9efa3293b59bfcb8d52c5fa2743306c94e399f5" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.302635 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee5d8876d1488183eb5edf35a9efa3293b59bfcb8d52c5fa2743306c94e399f5\": container with ID starting with ee5d8876d1488183eb5edf35a9efa3293b59bfcb8d52c5fa2743306c94e399f5 not found: ID does not exist" containerID="ee5d8876d1488183eb5edf35a9efa3293b59bfcb8d52c5fa2743306c94e399f5" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.302663 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee5d8876d1488183eb5edf35a9efa3293b59bfcb8d52c5fa2743306c94e399f5"} err="failed to get container status \"ee5d8876d1488183eb5edf35a9efa3293b59bfcb8d52c5fa2743306c94e399f5\": rpc error: code = NotFound desc = could not find container \"ee5d8876d1488183eb5edf35a9efa3293b59bfcb8d52c5fa2743306c94e399f5\": container with ID starting with ee5d8876d1488183eb5edf35a9efa3293b59bfcb8d52c5fa2743306c94e399f5 not found: ID does not exist" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.302682 4857 scope.go:117] "RemoveContainer" containerID="0d3f598c3d03bf8ce8f4bef7c23fdabe3795b3736cc7a62fc550d7a372dca74c" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.317911 4857 scope.go:117] "RemoveContainer" containerID="246cf699c95224893fcba082295e051b68ac17f47fec785c15e5741502bac507" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.335892 4857 scope.go:117] "RemoveContainer" containerID="696eb6717c35018debe86fc103767766044bac46fddb0ca57aa1307661843eda" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.362520 4857 scope.go:117] "RemoveContainer" containerID="0d3f598c3d03bf8ce8f4bef7c23fdabe3795b3736cc7a62fc550d7a372dca74c" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.363008 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d3f598c3d03bf8ce8f4bef7c23fdabe3795b3736cc7a62fc550d7a372dca74c\": container with ID starting with 0d3f598c3d03bf8ce8f4bef7c23fdabe3795b3736cc7a62fc550d7a372dca74c not found: ID does not exist" containerID="0d3f598c3d03bf8ce8f4bef7c23fdabe3795b3736cc7a62fc550d7a372dca74c" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.363060 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d3f598c3d03bf8ce8f4bef7c23fdabe3795b3736cc7a62fc550d7a372dca74c"} err="failed to get container status \"0d3f598c3d03bf8ce8f4bef7c23fdabe3795b3736cc7a62fc550d7a372dca74c\": rpc error: code = NotFound desc = could not find container \"0d3f598c3d03bf8ce8f4bef7c23fdabe3795b3736cc7a62fc550d7a372dca74c\": container with ID starting with 0d3f598c3d03bf8ce8f4bef7c23fdabe3795b3736cc7a62fc550d7a372dca74c not found: ID does not exist" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.363090 4857 scope.go:117] "RemoveContainer" containerID="246cf699c95224893fcba082295e051b68ac17f47fec785c15e5741502bac507" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.363422 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"246cf699c95224893fcba082295e051b68ac17f47fec785c15e5741502bac507\": container with ID starting with 246cf699c95224893fcba082295e051b68ac17f47fec785c15e5741502bac507 not found: ID does not exist" containerID="246cf699c95224893fcba082295e051b68ac17f47fec785c15e5741502bac507" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.363455 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"246cf699c95224893fcba082295e051b68ac17f47fec785c15e5741502bac507"} err="failed to get container status \"246cf699c95224893fcba082295e051b68ac17f47fec785c15e5741502bac507\": rpc error: code = NotFound desc = could not find container \"246cf699c95224893fcba082295e051b68ac17f47fec785c15e5741502bac507\": container with ID starting with 246cf699c95224893fcba082295e051b68ac17f47fec785c15e5741502bac507 not found: ID does not exist" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.363495 4857 scope.go:117] "RemoveContainer" containerID="696eb6717c35018debe86fc103767766044bac46fddb0ca57aa1307661843eda" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.363854 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"696eb6717c35018debe86fc103767766044bac46fddb0ca57aa1307661843eda\": container with ID starting with 696eb6717c35018debe86fc103767766044bac46fddb0ca57aa1307661843eda not found: ID does not exist" containerID="696eb6717c35018debe86fc103767766044bac46fddb0ca57aa1307661843eda" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.363878 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"696eb6717c35018debe86fc103767766044bac46fddb0ca57aa1307661843eda"} err="failed to get container status \"696eb6717c35018debe86fc103767766044bac46fddb0ca57aa1307661843eda\": rpc error: code = NotFound desc = could not find container \"696eb6717c35018debe86fc103767766044bac46fddb0ca57aa1307661843eda\": container with ID starting with 696eb6717c35018debe86fc103767766044bac46fddb0ca57aa1307661843eda not found: ID does not exist" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.363893 4857 scope.go:117] "RemoveContainer" containerID="1837b66f71bf7805f8b4307a67a79a46079d745062c0638df7ba39b0091df73b" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.368761 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d73807b-5e0c-4607-8393-5123a6cdbf9b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d73807b-5e0c-4607-8393-5123a6cdbf9b" (UID: "6d73807b-5e0c-4607-8393-5123a6cdbf9b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.387538 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d73807b-5e0c-4607-8393-5123a6cdbf9b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.399323 4857 scope.go:117] "RemoveContainer" containerID="1837b66f71bf7805f8b4307a67a79a46079d745062c0638df7ba39b0091df73b" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.399632 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1837b66f71bf7805f8b4307a67a79a46079d745062c0638df7ba39b0091df73b\": container with ID starting with 1837b66f71bf7805f8b4307a67a79a46079d745062c0638df7ba39b0091df73b not found: ID does not exist" containerID="1837b66f71bf7805f8b4307a67a79a46079d745062c0638df7ba39b0091df73b" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.399657 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1837b66f71bf7805f8b4307a67a79a46079d745062c0638df7ba39b0091df73b"} err="failed to get container status \"1837b66f71bf7805f8b4307a67a79a46079d745062c0638df7ba39b0091df73b\": rpc error: code = NotFound desc = could not find container \"1837b66f71bf7805f8b4307a67a79a46079d745062c0638df7ba39b0091df73b\": container with ID starting with 1837b66f71bf7805f8b4307a67a79a46079d745062c0638df7ba39b0091df73b not found: ID does not exist" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.399677 4857 scope.go:117] "RemoveContainer" containerID="34c2234b5cda4e087d988f3152f4a10fc71a296bec05bfffd187d84e66ce1cf5" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.415858 4857 scope.go:117] "RemoveContainer" containerID="c75193c95e07cc6573aa6fd4397e64e48dba1f87c3b51cf693866eb540a1f56a" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.462140 4857 log.go:32] "RunPodSandbox from runtime service failed" err=< Feb 22 00:10:30 crc kubenswrapper[4857]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-4cnjs_openshift-marketplace_aec64a2e-ff83-4d35-a6cc-691c906ee0d9_0(5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e): error adding pod openshift-marketplace_marketplace-operator-79b997595-4cnjs to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e" Netns:"/var/run/netns/450408da-461f-47ea-bcb5-86b06d6999d9" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-4cnjs;K8S_POD_INFRA_CONTAINER_ID=5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e;K8S_POD_UID=aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-4cnjs] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-4cnjs/aec64a2e-ff83-4d35-a6cc-691c906ee0d9]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4cnjs?timeout=1m0s": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:10:30 crc kubenswrapper[4857]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 22 00:10:30 crc kubenswrapper[4857]: > Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.462227 4857 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Feb 22 00:10:30 crc kubenswrapper[4857]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-4cnjs_openshift-marketplace_aec64a2e-ff83-4d35-a6cc-691c906ee0d9_0(5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e): error adding pod openshift-marketplace_marketplace-operator-79b997595-4cnjs to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e" Netns:"/var/run/netns/450408da-461f-47ea-bcb5-86b06d6999d9" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-4cnjs;K8S_POD_INFRA_CONTAINER_ID=5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e;K8S_POD_UID=aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-4cnjs] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-4cnjs/aec64a2e-ff83-4d35-a6cc-691c906ee0d9]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4cnjs?timeout=1m0s": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:10:30 crc kubenswrapper[4857]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 22 00:10:30 crc kubenswrapper[4857]: > pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.462247 4857 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Feb 22 00:10:30 crc kubenswrapper[4857]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-4cnjs_openshift-marketplace_aec64a2e-ff83-4d35-a6cc-691c906ee0d9_0(5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e): error adding pod openshift-marketplace_marketplace-operator-79b997595-4cnjs to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e" Netns:"/var/run/netns/450408da-461f-47ea-bcb5-86b06d6999d9" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-4cnjs;K8S_POD_INFRA_CONTAINER_ID=5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e;K8S_POD_UID=aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-4cnjs] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-4cnjs/aec64a2e-ff83-4d35-a6cc-691c906ee0d9]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4cnjs?timeout=1m0s": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:10:30 crc kubenswrapper[4857]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 22 00:10:30 crc kubenswrapper[4857]: > pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.462310 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"marketplace-operator-79b997595-4cnjs_openshift-marketplace(aec64a2e-ff83-4d35-a6cc-691c906ee0d9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"marketplace-operator-79b997595-4cnjs_openshift-marketplace(aec64a2e-ff83-4d35-a6cc-691c906ee0d9)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-4cnjs_openshift-marketplace_aec64a2e-ff83-4d35-a6cc-691c906ee0d9_0(5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e): error adding pod openshift-marketplace_marketplace-operator-79b997595-4cnjs to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e\\\" Netns:\\\"/var/run/netns/450408da-461f-47ea-bcb5-86b06d6999d9\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-4cnjs;K8S_POD_INFRA_CONTAINER_ID=5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e;K8S_POD_UID=aec64a2e-ff83-4d35-a6cc-691c906ee0d9\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-4cnjs] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-4cnjs/aec64a2e-ff83-4d35-a6cc-691c906ee0d9]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: status update failed for pod /: Get \\\"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4cnjs?timeout=1m0s\\\": dial tcp 38.102.83.180:6443: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" podUID="aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.462765 4857 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.180:6443: connect: connection refused" event=< Feb 22 00:10:30 crc kubenswrapper[4857]: &Event{ObjectMeta:{marketplace-operator-79b997595-4cnjs.189669123c80c465 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:marketplace-operator-79b997595-4cnjs,UID:aec64a2e-ff83-4d35-a6cc-691c906ee0d9,APIVersion:v1,ResourceVersion:29623,FieldPath:,},Reason:FailedCreatePodSandBox,Message:Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-4cnjs_openshift-marketplace_aec64a2e-ff83-4d35-a6cc-691c906ee0d9_0(5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e): error adding pod openshift-marketplace_marketplace-operator-79b997595-4cnjs to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e" Netns:"/var/run/netns/450408da-461f-47ea-bcb5-86b06d6999d9" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-4cnjs;K8S_POD_INFRA_CONTAINER_ID=5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e;K8S_POD_UID=aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-4cnjs] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-4cnjs/aec64a2e-ff83-4d35-a6cc-691c906ee0d9]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4cnjs?timeout=1m0s": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:10:30 crc kubenswrapper[4857]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"},Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-22 00:10:30.462268517 +0000 UTC m=+238.100997770,LastTimestamp:2026-02-22 00:10:30.462268517 +0000 UTC m=+238.100997770,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 22 00:10:30 crc kubenswrapper[4857]: > Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.486123 4857 scope.go:117] "RemoveContainer" containerID="4ab98083305747c64255ef608e22c65a22fd727e68fb36f8611ab725e7b14ba4" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.514901 4857 scope.go:117] "RemoveContainer" containerID="34c2234b5cda4e087d988f3152f4a10fc71a296bec05bfffd187d84e66ce1cf5" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.515452 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34c2234b5cda4e087d988f3152f4a10fc71a296bec05bfffd187d84e66ce1cf5\": container with ID starting with 34c2234b5cda4e087d988f3152f4a10fc71a296bec05bfffd187d84e66ce1cf5 not found: ID does not exist" containerID="34c2234b5cda4e087d988f3152f4a10fc71a296bec05bfffd187d84e66ce1cf5" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.515491 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34c2234b5cda4e087d988f3152f4a10fc71a296bec05bfffd187d84e66ce1cf5"} err="failed to get container status \"34c2234b5cda4e087d988f3152f4a10fc71a296bec05bfffd187d84e66ce1cf5\": rpc error: code = NotFound desc = could not find container \"34c2234b5cda4e087d988f3152f4a10fc71a296bec05bfffd187d84e66ce1cf5\": container with ID starting with 34c2234b5cda4e087d988f3152f4a10fc71a296bec05bfffd187d84e66ce1cf5 not found: ID does not exist" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.515517 4857 scope.go:117] "RemoveContainer" containerID="c75193c95e07cc6573aa6fd4397e64e48dba1f87c3b51cf693866eb540a1f56a" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.516094 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c75193c95e07cc6573aa6fd4397e64e48dba1f87c3b51cf693866eb540a1f56a\": container with ID starting with c75193c95e07cc6573aa6fd4397e64e48dba1f87c3b51cf693866eb540a1f56a not found: ID does not exist" containerID="c75193c95e07cc6573aa6fd4397e64e48dba1f87c3b51cf693866eb540a1f56a" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.516185 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c75193c95e07cc6573aa6fd4397e64e48dba1f87c3b51cf693866eb540a1f56a"} err="failed to get container status \"c75193c95e07cc6573aa6fd4397e64e48dba1f87c3b51cf693866eb540a1f56a\": rpc error: code = NotFound desc = could not find container \"c75193c95e07cc6573aa6fd4397e64e48dba1f87c3b51cf693866eb540a1f56a\": container with ID starting with c75193c95e07cc6573aa6fd4397e64e48dba1f87c3b51cf693866eb540a1f56a not found: ID does not exist" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.516264 4857 scope.go:117] "RemoveContainer" containerID="4ab98083305747c64255ef608e22c65a22fd727e68fb36f8611ab725e7b14ba4" Feb 22 00:10:30 crc kubenswrapper[4857]: E0222 00:10:30.516835 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ab98083305747c64255ef608e22c65a22fd727e68fb36f8611ab725e7b14ba4\": container with ID starting with 4ab98083305747c64255ef608e22c65a22fd727e68fb36f8611ab725e7b14ba4 not found: ID does not exist" containerID="4ab98083305747c64255ef608e22c65a22fd727e68fb36f8611ab725e7b14ba4" Feb 22 00:10:30 crc kubenswrapper[4857]: I0222 00:10:30.516890 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ab98083305747c64255ef608e22c65a22fd727e68fb36f8611ab725e7b14ba4"} err="failed to get container status \"4ab98083305747c64255ef608e22c65a22fd727e68fb36f8611ab725e7b14ba4\": rpc error: code = NotFound desc = could not find container \"4ab98083305747c64255ef608e22c65a22fd727e68fb36f8611ab725e7b14ba4\": container with ID starting with 4ab98083305747c64255ef608e22c65a22fd727e68fb36f8611ab725e7b14ba4 not found: ID does not exist" Feb 22 00:10:31 crc kubenswrapper[4857]: I0222 00:10:31.171760 4857 generic.go:334] "Generic (PLEG): container finished" podID="7ab5d710-d809-4d86-b56b-1f611cb02644" containerID="3694ae29f8d75fe25793827365d4a12165d51adc31c47f3273747ed2a84bdfa4" exitCode=0 Feb 22 00:10:31 crc kubenswrapper[4857]: I0222 00:10:31.171818 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7ab5d710-d809-4d86-b56b-1f611cb02644","Type":"ContainerDied","Data":"3694ae29f8d75fe25793827365d4a12165d51adc31c47f3273747ed2a84bdfa4"} Feb 22 00:10:31 crc kubenswrapper[4857]: I0222 00:10:31.176421 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 22 00:10:31 crc kubenswrapper[4857]: I0222 00:10:31.177546 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 22 00:10:31 crc kubenswrapper[4857]: I0222 00:10:31.178241 4857 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918" exitCode=0 Feb 22 00:10:31 crc kubenswrapper[4857]: I0222 00:10:31.178272 4857 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa" exitCode=0 Feb 22 00:10:31 crc kubenswrapper[4857]: I0222 00:10:31.178279 4857 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f" exitCode=0 Feb 22 00:10:31 crc kubenswrapper[4857]: I0222 00:10:31.178286 4857 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45" exitCode=2 Feb 22 00:10:31 crc kubenswrapper[4857]: I0222 00:10:31.178302 4857 scope.go:117] "RemoveContainer" containerID="190ca09af92f2f5888e1310679c4e3fc1ad39aa0816fbd88ada6a5c445443253" Feb 22 00:10:31 crc kubenswrapper[4857]: I0222 00:10:31.178330 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:31 crc kubenswrapper[4857]: I0222 00:10:31.178704 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:31 crc kubenswrapper[4857]: E0222 00:10:31.494853 4857 log.go:32] "RunPodSandbox from runtime service failed" err=< Feb 22 00:10:31 crc kubenswrapper[4857]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-4cnjs_openshift-marketplace_aec64a2e-ff83-4d35-a6cc-691c906ee0d9_0(53389f9d859c63bd35cfd970ef9b3ad7a66d1e008311393c82f176c4104b543a): error adding pod openshift-marketplace_marketplace-operator-79b997595-4cnjs to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"53389f9d859c63bd35cfd970ef9b3ad7a66d1e008311393c82f176c4104b543a" Netns:"/var/run/netns/a7d2d58a-2afc-4bc9-8bc8-b936dd5dfd08" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-4cnjs;K8S_POD_INFRA_CONTAINER_ID=53389f9d859c63bd35cfd970ef9b3ad7a66d1e008311393c82f176c4104b543a;K8S_POD_UID=aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-4cnjs] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-4cnjs/aec64a2e-ff83-4d35-a6cc-691c906ee0d9]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4cnjs?timeout=1m0s": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:10:31 crc kubenswrapper[4857]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 22 00:10:31 crc kubenswrapper[4857]: > Feb 22 00:10:31 crc kubenswrapper[4857]: E0222 00:10:31.495294 4857 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Feb 22 00:10:31 crc kubenswrapper[4857]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-4cnjs_openshift-marketplace_aec64a2e-ff83-4d35-a6cc-691c906ee0d9_0(53389f9d859c63bd35cfd970ef9b3ad7a66d1e008311393c82f176c4104b543a): error adding pod openshift-marketplace_marketplace-operator-79b997595-4cnjs to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"53389f9d859c63bd35cfd970ef9b3ad7a66d1e008311393c82f176c4104b543a" Netns:"/var/run/netns/a7d2d58a-2afc-4bc9-8bc8-b936dd5dfd08" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-4cnjs;K8S_POD_INFRA_CONTAINER_ID=53389f9d859c63bd35cfd970ef9b3ad7a66d1e008311393c82f176c4104b543a;K8S_POD_UID=aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-4cnjs] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-4cnjs/aec64a2e-ff83-4d35-a6cc-691c906ee0d9]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4cnjs?timeout=1m0s": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:10:31 crc kubenswrapper[4857]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 22 00:10:31 crc kubenswrapper[4857]: > pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:31 crc kubenswrapper[4857]: E0222 00:10:31.495337 4857 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Feb 22 00:10:31 crc kubenswrapper[4857]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-4cnjs_openshift-marketplace_aec64a2e-ff83-4d35-a6cc-691c906ee0d9_0(53389f9d859c63bd35cfd970ef9b3ad7a66d1e008311393c82f176c4104b543a): error adding pod openshift-marketplace_marketplace-operator-79b997595-4cnjs to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"53389f9d859c63bd35cfd970ef9b3ad7a66d1e008311393c82f176c4104b543a" Netns:"/var/run/netns/a7d2d58a-2afc-4bc9-8bc8-b936dd5dfd08" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-4cnjs;K8S_POD_INFRA_CONTAINER_ID=53389f9d859c63bd35cfd970ef9b3ad7a66d1e008311393c82f176c4104b543a;K8S_POD_UID=aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-4cnjs] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-4cnjs/aec64a2e-ff83-4d35-a6cc-691c906ee0d9]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4cnjs?timeout=1m0s": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:10:31 crc kubenswrapper[4857]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 22 00:10:31 crc kubenswrapper[4857]: > pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:31 crc kubenswrapper[4857]: E0222 00:10:31.495410 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"marketplace-operator-79b997595-4cnjs_openshift-marketplace(aec64a2e-ff83-4d35-a6cc-691c906ee0d9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"marketplace-operator-79b997595-4cnjs_openshift-marketplace(aec64a2e-ff83-4d35-a6cc-691c906ee0d9)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-4cnjs_openshift-marketplace_aec64a2e-ff83-4d35-a6cc-691c906ee0d9_0(53389f9d859c63bd35cfd970ef9b3ad7a66d1e008311393c82f176c4104b543a): error adding pod openshift-marketplace_marketplace-operator-79b997595-4cnjs to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"53389f9d859c63bd35cfd970ef9b3ad7a66d1e008311393c82f176c4104b543a\\\" Netns:\\\"/var/run/netns/a7d2d58a-2afc-4bc9-8bc8-b936dd5dfd08\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-4cnjs;K8S_POD_INFRA_CONTAINER_ID=53389f9d859c63bd35cfd970ef9b3ad7a66d1e008311393c82f176c4104b543a;K8S_POD_UID=aec64a2e-ff83-4d35-a6cc-691c906ee0d9\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-4cnjs] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-4cnjs/aec64a2e-ff83-4d35-a6cc-691c906ee0d9]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: status update failed for pod /: Get \\\"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4cnjs?timeout=1m0s\\\": dial tcp 38.102.83.180:6443: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" podUID="aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Feb 22 00:10:31 crc kubenswrapper[4857]: E0222 00:10:31.981731 4857 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.180:6443: connect: connection refused" event=< Feb 22 00:10:31 crc kubenswrapper[4857]: &Event{ObjectMeta:{marketplace-operator-79b997595-4cnjs.189669123c80c465 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:marketplace-operator-79b997595-4cnjs,UID:aec64a2e-ff83-4d35-a6cc-691c906ee0d9,APIVersion:v1,ResourceVersion:29623,FieldPath:,},Reason:FailedCreatePodSandBox,Message:Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-4cnjs_openshift-marketplace_aec64a2e-ff83-4d35-a6cc-691c906ee0d9_0(5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e): error adding pod openshift-marketplace_marketplace-operator-79b997595-4cnjs to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e" Netns:"/var/run/netns/450408da-461f-47ea-bcb5-86b06d6999d9" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-4cnjs;K8S_POD_INFRA_CONTAINER_ID=5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e;K8S_POD_UID=aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-4cnjs] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-4cnjs/aec64a2e-ff83-4d35-a6cc-691c906ee0d9]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4cnjs?timeout=1m0s": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:10:31 crc kubenswrapper[4857]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"},Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-22 00:10:30.462268517 +0000 UTC m=+238.100997770,LastTimestamp:2026-02-22 00:10:30.462268517 +0000 UTC m=+238.100997770,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 22 00:10:31 crc kubenswrapper[4857]: > Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.186798 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.403953 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.405162 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.407195 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.516846 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7ab5d710-d809-4d86-b56b-1f611cb02644-var-lock\") pod \"7ab5d710-d809-4d86-b56b-1f611cb02644\" (UID: \"7ab5d710-d809-4d86-b56b-1f611cb02644\") " Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.516896 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.516920 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ab5d710-d809-4d86-b56b-1f611cb02644-kubelet-dir\") pod \"7ab5d710-d809-4d86-b56b-1f611cb02644\" (UID: \"7ab5d710-d809-4d86-b56b-1f611cb02644\") " Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.516987 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ab5d710-d809-4d86-b56b-1f611cb02644-kube-api-access\") pod \"7ab5d710-d809-4d86-b56b-1f611cb02644\" (UID: \"7ab5d710-d809-4d86-b56b-1f611cb02644\") " Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.516980 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ab5d710-d809-4d86-b56b-1f611cb02644-var-lock" (OuterVolumeSpecName: "var-lock") pod "7ab5d710-d809-4d86-b56b-1f611cb02644" (UID: "7ab5d710-d809-4d86-b56b-1f611cb02644"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.517100 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.517011 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.517057 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ab5d710-d809-4d86-b56b-1f611cb02644-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7ab5d710-d809-4d86-b56b-1f611cb02644" (UID: "7ab5d710-d809-4d86-b56b-1f611cb02644"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.517083 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.517167 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.517251 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.517566 4857 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.517581 4857 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.517589 4857 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7ab5d710-d809-4d86-b56b-1f611cb02644-var-lock\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.517598 4857 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.517606 4857 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ab5d710-d809-4d86-b56b-1f611cb02644-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.522291 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ab5d710-d809-4d86-b56b-1f611cb02644-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7ab5d710-d809-4d86-b56b-1f611cb02644" (UID: "7ab5d710-d809-4d86-b56b-1f611cb02644"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:10:32 crc kubenswrapper[4857]: I0222 00:10:32.619319 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ab5d710-d809-4d86-b56b-1f611cb02644-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.085249 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.194605 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.194877 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7ab5d710-d809-4d86-b56b-1f611cb02644","Type":"ContainerDied","Data":"7bf51deb2895b48144471757f30daa5e82610e4d537ea4fa71ae9d2edda4f000"} Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.194919 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bf51deb2895b48144471757f30daa5e82610e4d537ea4fa71ae9d2edda4f000" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.200186 4857 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df" exitCode=0 Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.200244 4857 scope.go:117] "RemoveContainer" containerID="e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.200756 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.219821 4857 scope.go:117] "RemoveContainer" containerID="1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.233618 4857 scope.go:117] "RemoveContainer" containerID="ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.244688 4857 scope.go:117] "RemoveContainer" containerID="f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.254606 4857 scope.go:117] "RemoveContainer" containerID="af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.269283 4857 scope.go:117] "RemoveContainer" containerID="e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.288011 4857 scope.go:117] "RemoveContainer" containerID="e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918" Feb 22 00:10:33 crc kubenswrapper[4857]: E0222 00:10:33.288478 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\": container with ID starting with e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918 not found: ID does not exist" containerID="e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.288527 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918"} err="failed to get container status \"e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\": rpc error: code = NotFound desc = could not find container \"e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918\": container with ID starting with e7d09936de1d8a8c2ac7a8dc323ebebe1133f801b3b7852e1bb2ba0b7dea4918 not found: ID does not exist" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.288560 4857 scope.go:117] "RemoveContainer" containerID="1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa" Feb 22 00:10:33 crc kubenswrapper[4857]: E0222 00:10:33.288837 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\": container with ID starting with 1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa not found: ID does not exist" containerID="1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.288874 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa"} err="failed to get container status \"1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\": rpc error: code = NotFound desc = could not find container \"1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa\": container with ID starting with 1cac3bbd0bebf2830a8d99058b27e5b7ffbfcba252ac89f3d5c691aacd875baa not found: ID does not exist" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.288902 4857 scope.go:117] "RemoveContainer" containerID="ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f" Feb 22 00:10:33 crc kubenswrapper[4857]: E0222 00:10:33.289164 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\": container with ID starting with ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f not found: ID does not exist" containerID="ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.289192 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f"} err="failed to get container status \"ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\": rpc error: code = NotFound desc = could not find container \"ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f\": container with ID starting with ac43be4d4e931967d20574c433c9c2e4acea56f7a8fb72406ae1bdf10607096f not found: ID does not exist" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.289244 4857 scope.go:117] "RemoveContainer" containerID="f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45" Feb 22 00:10:33 crc kubenswrapper[4857]: E0222 00:10:33.289533 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\": container with ID starting with f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45 not found: ID does not exist" containerID="f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.289572 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45"} err="failed to get container status \"f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\": rpc error: code = NotFound desc = could not find container \"f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45\": container with ID starting with f78e5872e55a9c51fde7028b883c3946872ab7a0eee7eb5f86eb9a19ee42cf45 not found: ID does not exist" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.289599 4857 scope.go:117] "RemoveContainer" containerID="af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df" Feb 22 00:10:33 crc kubenswrapper[4857]: E0222 00:10:33.290088 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\": container with ID starting with af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df not found: ID does not exist" containerID="af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.290116 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df"} err="failed to get container status \"af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\": rpc error: code = NotFound desc = could not find container \"af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df\": container with ID starting with af4b250c27a443693739e4694ea980b7cec7adb57f00c7dd143a369a02a222df not found: ID does not exist" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.290142 4857 scope.go:117] "RemoveContainer" containerID="e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f" Feb 22 00:10:33 crc kubenswrapper[4857]: E0222 00:10:33.290416 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\": container with ID starting with e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f not found: ID does not exist" containerID="e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.290482 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f"} err="failed to get container status \"e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\": rpc error: code = NotFound desc = could not find container \"e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f\": container with ID starting with e8ac9788aea9e427e98f55eb810950fae6579e54265e430e3f4a3b382e502c0f not found: ID does not exist" Feb 22 00:10:33 crc kubenswrapper[4857]: E0222 00:10:33.532560 4857 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:33 crc kubenswrapper[4857]: E0222 00:10:33.532968 4857 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:33 crc kubenswrapper[4857]: E0222 00:10:33.533255 4857 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:33 crc kubenswrapper[4857]: E0222 00:10:33.533412 4857 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:33 crc kubenswrapper[4857]: E0222 00:10:33.533555 4857 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:33 crc kubenswrapper[4857]: I0222 00:10:33.533579 4857 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 22 00:10:33 crc kubenswrapper[4857]: E0222 00:10:33.533709 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="200ms" Feb 22 00:10:33 crc kubenswrapper[4857]: E0222 00:10:33.735028 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="400ms" Feb 22 00:10:34 crc kubenswrapper[4857]: E0222 00:10:34.136396 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="800ms" Feb 22 00:10:34 crc kubenswrapper[4857]: E0222 00:10:34.936810 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="1.6s" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.028118 4857 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.028370 4857 status_manager.go:851] "Failed to get status for pod" podUID="4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-vmk2g\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.028539 4857 status_manager.go:851] "Failed to get status for pod" podUID="10f2d337-12f1-4949-9872-08df9aa084ab" pod="openshift-marketplace/redhat-marketplace-797dx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-797dx\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.028690 4857 status_manager.go:851] "Failed to get status for pod" podUID="18611d84-22b3-49df-9569-ac383eb42cab" pod="openshift-marketplace/certified-operators-m5rft" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-m5rft\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.031482 4857 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.031639 4857 status_manager.go:851] "Failed to get status for pod" podUID="4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-vmk2g\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.031781 4857 status_manager.go:851] "Failed to get status for pod" podUID="7ab5d710-d809-4d86-b56b-1f611cb02644" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.031920 4857 status_manager.go:851] "Failed to get status for pod" podUID="fa45663f-d64c-431c-be1d-29175cc804ea" pod="openshift-marketplace/community-operators-5rpdr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-5rpdr\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.032112 4857 status_manager.go:851] "Failed to get status for pod" podUID="10f2d337-12f1-4949-9872-08df9aa084ab" pod="openshift-marketplace/redhat-marketplace-797dx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-797dx\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.032249 4857 status_manager.go:851] "Failed to get status for pod" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" pod="openshift-marketplace/redhat-operators-xd4g7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xd4g7\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.032384 4857 status_manager.go:851] "Failed to get status for pod" podUID="18611d84-22b3-49df-9569-ac383eb42cab" pod="openshift-marketplace/certified-operators-m5rft" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-m5rft\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.032554 4857 status_manager.go:851] "Failed to get status for pod" podUID="10f2d337-12f1-4949-9872-08df9aa084ab" pod="openshift-marketplace/redhat-marketplace-797dx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-797dx\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.032685 4857 status_manager.go:851] "Failed to get status for pod" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" pod="openshift-marketplace/redhat-operators-xd4g7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xd4g7\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.032824 4857 status_manager.go:851] "Failed to get status for pod" podUID="18611d84-22b3-49df-9569-ac383eb42cab" pod="openshift-marketplace/certified-operators-m5rft" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-m5rft\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.032962 4857 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.033124 4857 status_manager.go:851] "Failed to get status for pod" podUID="4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-vmk2g\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.033270 4857 status_manager.go:851] "Failed to get status for pod" podUID="7ab5d710-d809-4d86-b56b-1f611cb02644" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.033411 4857 status_manager.go:851] "Failed to get status for pod" podUID="fa45663f-d64c-431c-be1d-29175cc804ea" pod="openshift-marketplace/community-operators-5rpdr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-5rpdr\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:35 crc kubenswrapper[4857]: E0222 00:10:35.066176 4857 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.180:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.066549 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:35 crc kubenswrapper[4857]: I0222 00:10:35.211355 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"0b37c92992e710d805b07ad9ffa78a5f9d7a8a756336f45750c9aee109e316b7"} Feb 22 00:10:36 crc kubenswrapper[4857]: I0222 00:10:36.217449 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"6f772d5bf4afd3597656f75b78ddd28f1cc5111bfbe4e0056153e48b09206633"} Feb 22 00:10:36 crc kubenswrapper[4857]: I0222 00:10:36.218164 4857 status_manager.go:851] "Failed to get status for pod" podUID="18611d84-22b3-49df-9569-ac383eb42cab" pod="openshift-marketplace/certified-operators-m5rft" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-m5rft\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:36 crc kubenswrapper[4857]: E0222 00:10:36.218273 4857 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.180:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:36 crc kubenswrapper[4857]: I0222 00:10:36.218444 4857 status_manager.go:851] "Failed to get status for pod" podUID="4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-vmk2g\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:36 crc kubenswrapper[4857]: I0222 00:10:36.218822 4857 status_manager.go:851] "Failed to get status for pod" podUID="7ab5d710-d809-4d86-b56b-1f611cb02644" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:36 crc kubenswrapper[4857]: I0222 00:10:36.219120 4857 status_manager.go:851] "Failed to get status for pod" podUID="fa45663f-d64c-431c-be1d-29175cc804ea" pod="openshift-marketplace/community-operators-5rpdr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-5rpdr\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:36 crc kubenswrapper[4857]: I0222 00:10:36.219348 4857 status_manager.go:851] "Failed to get status for pod" podUID="10f2d337-12f1-4949-9872-08df9aa084ab" pod="openshift-marketplace/redhat-marketplace-797dx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-797dx\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:36 crc kubenswrapper[4857]: I0222 00:10:36.219697 4857 status_manager.go:851] "Failed to get status for pod" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" pod="openshift-marketplace/redhat-operators-xd4g7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xd4g7\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:36 crc kubenswrapper[4857]: E0222 00:10:36.537367 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="3.2s" Feb 22 00:10:37 crc kubenswrapper[4857]: E0222 00:10:37.225303 4857 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.180:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:10:39 crc kubenswrapper[4857]: E0222 00:10:39.739344 4857 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="6.4s" Feb 22 00:10:41 crc kubenswrapper[4857]: E0222 00:10:41.983433 4857 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.180:6443: connect: connection refused" event=< Feb 22 00:10:41 crc kubenswrapper[4857]: &Event{ObjectMeta:{marketplace-operator-79b997595-4cnjs.189669123c80c465 openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:marketplace-operator-79b997595-4cnjs,UID:aec64a2e-ff83-4d35-a6cc-691c906ee0d9,APIVersion:v1,ResourceVersion:29623,FieldPath:,},Reason:FailedCreatePodSandBox,Message:Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-4cnjs_openshift-marketplace_aec64a2e-ff83-4d35-a6cc-691c906ee0d9_0(5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e): error adding pod openshift-marketplace_marketplace-operator-79b997595-4cnjs to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e" Netns:"/var/run/netns/450408da-461f-47ea-bcb5-86b06d6999d9" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-4cnjs;K8S_POD_INFRA_CONTAINER_ID=5bc9f5873b16ffd53a6fd34fdb5c4fd02575a192e1f6a40153ab169add66a71e;K8S_POD_UID=aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-4cnjs] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-4cnjs/aec64a2e-ff83-4d35-a6cc-691c906ee0d9]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4cnjs?timeout=1m0s": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:10:41 crc kubenswrapper[4857]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"},Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-22 00:10:30.462268517 +0000 UTC m=+238.100997770,LastTimestamp:2026-02-22 00:10:30.462268517 +0000 UTC m=+238.100997770,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 22 00:10:41 crc kubenswrapper[4857]: > Feb 22 00:10:43 crc kubenswrapper[4857]: I0222 00:10:43.081738 4857 status_manager.go:851] "Failed to get status for pod" podUID="fa45663f-d64c-431c-be1d-29175cc804ea" pod="openshift-marketplace/community-operators-5rpdr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-5rpdr\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:43 crc kubenswrapper[4857]: I0222 00:10:43.082446 4857 status_manager.go:851] "Failed to get status for pod" podUID="10f2d337-12f1-4949-9872-08df9aa084ab" pod="openshift-marketplace/redhat-marketplace-797dx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-797dx\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:43 crc kubenswrapper[4857]: I0222 00:10:43.082993 4857 status_manager.go:851] "Failed to get status for pod" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" pod="openshift-marketplace/redhat-operators-xd4g7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xd4g7\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:43 crc kubenswrapper[4857]: I0222 00:10:43.083443 4857 status_manager.go:851] "Failed to get status for pod" podUID="18611d84-22b3-49df-9569-ac383eb42cab" pod="openshift-marketplace/certified-operators-m5rft" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-m5rft\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:43 crc kubenswrapper[4857]: I0222 00:10:43.083840 4857 status_manager.go:851] "Failed to get status for pod" podUID="4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-vmk2g\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:43 crc kubenswrapper[4857]: I0222 00:10:43.084320 4857 status_manager.go:851] "Failed to get status for pod" podUID="7ab5d710-d809-4d86-b56b-1f611cb02644" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:43 crc kubenswrapper[4857]: E0222 00:10:43.151492 4857 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.180:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-96hns" volumeName="registry-storage" Feb 22 00:10:44 crc kubenswrapper[4857]: I0222 00:10:44.077505 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:44 crc kubenswrapper[4857]: I0222 00:10:44.077545 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:44 crc kubenswrapper[4857]: I0222 00:10:44.078492 4857 status_manager.go:851] "Failed to get status for pod" podUID="fa45663f-d64c-431c-be1d-29175cc804ea" pod="openshift-marketplace/community-operators-5rpdr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-5rpdr\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:44 crc kubenswrapper[4857]: I0222 00:10:44.078717 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:44 crc kubenswrapper[4857]: I0222 00:10:44.082049 4857 status_manager.go:851] "Failed to get status for pod" podUID="10f2d337-12f1-4949-9872-08df9aa084ab" pod="openshift-marketplace/redhat-marketplace-797dx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-797dx\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:44 crc kubenswrapper[4857]: I0222 00:10:44.082494 4857 status_manager.go:851] "Failed to get status for pod" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" pod="openshift-marketplace/redhat-operators-xd4g7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xd4g7\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:44 crc kubenswrapper[4857]: I0222 00:10:44.083209 4857 status_manager.go:851] "Failed to get status for pod" podUID="18611d84-22b3-49df-9569-ac383eb42cab" pod="openshift-marketplace/certified-operators-m5rft" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-m5rft\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:44 crc kubenswrapper[4857]: I0222 00:10:44.083662 4857 status_manager.go:851] "Failed to get status for pod" podUID="4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-vmk2g\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:44 crc kubenswrapper[4857]: I0222 00:10:44.084009 4857 status_manager.go:851] "Failed to get status for pod" podUID="7ab5d710-d809-4d86-b56b-1f611cb02644" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:44 crc kubenswrapper[4857]: I0222 00:10:44.159309 4857 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8bfed332-8a77-48dd-b3c6-7068b918dc71" Feb 22 00:10:44 crc kubenswrapper[4857]: I0222 00:10:44.159350 4857 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8bfed332-8a77-48dd-b3c6-7068b918dc71" Feb 22 00:10:44 crc kubenswrapper[4857]: E0222 00:10:44.159799 4857 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:44 crc kubenswrapper[4857]: I0222 00:10:44.160880 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:44 crc kubenswrapper[4857]: I0222 00:10:44.266310 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f47fc77ae071207eed1d61e85c841f3a715fc569f048bcfdb954816346263b08"} Feb 22 00:10:44 crc kubenswrapper[4857]: E0222 00:10:44.455774 4857 log.go:32] "RunPodSandbox from runtime service failed" err=< Feb 22 00:10:44 crc kubenswrapper[4857]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-4cnjs_openshift-marketplace_aec64a2e-ff83-4d35-a6cc-691c906ee0d9_0(ef18c54c47c7b3c34a22a9810c2da42d79dc33698d658fb0fe0e34254be408ac): error adding pod openshift-marketplace_marketplace-operator-79b997595-4cnjs to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ef18c54c47c7b3c34a22a9810c2da42d79dc33698d658fb0fe0e34254be408ac" Netns:"/var/run/netns/a5a51a98-9d23-4650-bbb4-f21904646377" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-4cnjs;K8S_POD_INFRA_CONTAINER_ID=ef18c54c47c7b3c34a22a9810c2da42d79dc33698d658fb0fe0e34254be408ac;K8S_POD_UID=aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-4cnjs] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-4cnjs/aec64a2e-ff83-4d35-a6cc-691c906ee0d9]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4cnjs?timeout=1m0s": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:10:44 crc kubenswrapper[4857]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 22 00:10:44 crc kubenswrapper[4857]: > Feb 22 00:10:44 crc kubenswrapper[4857]: E0222 00:10:44.455853 4857 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Feb 22 00:10:44 crc kubenswrapper[4857]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-4cnjs_openshift-marketplace_aec64a2e-ff83-4d35-a6cc-691c906ee0d9_0(ef18c54c47c7b3c34a22a9810c2da42d79dc33698d658fb0fe0e34254be408ac): error adding pod openshift-marketplace_marketplace-operator-79b997595-4cnjs to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ef18c54c47c7b3c34a22a9810c2da42d79dc33698d658fb0fe0e34254be408ac" Netns:"/var/run/netns/a5a51a98-9d23-4650-bbb4-f21904646377" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-4cnjs;K8S_POD_INFRA_CONTAINER_ID=ef18c54c47c7b3c34a22a9810c2da42d79dc33698d658fb0fe0e34254be408ac;K8S_POD_UID=aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-4cnjs] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-4cnjs/aec64a2e-ff83-4d35-a6cc-691c906ee0d9]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4cnjs?timeout=1m0s": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:10:44 crc kubenswrapper[4857]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 22 00:10:44 crc kubenswrapper[4857]: > pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:44 crc kubenswrapper[4857]: E0222 00:10:44.455878 4857 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Feb 22 00:10:44 crc kubenswrapper[4857]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-4cnjs_openshift-marketplace_aec64a2e-ff83-4d35-a6cc-691c906ee0d9_0(ef18c54c47c7b3c34a22a9810c2da42d79dc33698d658fb0fe0e34254be408ac): error adding pod openshift-marketplace_marketplace-operator-79b997595-4cnjs to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ef18c54c47c7b3c34a22a9810c2da42d79dc33698d658fb0fe0e34254be408ac" Netns:"/var/run/netns/a5a51a98-9d23-4650-bbb4-f21904646377" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-4cnjs;K8S_POD_INFRA_CONTAINER_ID=ef18c54c47c7b3c34a22a9810c2da42d79dc33698d658fb0fe0e34254be408ac;K8S_POD_UID=aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Path:"" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-4cnjs] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-4cnjs/aec64a2e-ff83-4d35-a6cc-691c906ee0d9]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: status update failed for pod /: Get "https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4cnjs?timeout=1m0s": dial tcp 38.102.83.180:6443: connect: connection refused Feb 22 00:10:44 crc kubenswrapper[4857]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Feb 22 00:10:44 crc kubenswrapper[4857]: > pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:44 crc kubenswrapper[4857]: E0222 00:10:44.455954 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"marketplace-operator-79b997595-4cnjs_openshift-marketplace(aec64a2e-ff83-4d35-a6cc-691c906ee0d9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"marketplace-operator-79b997595-4cnjs_openshift-marketplace(aec64a2e-ff83-4d35-a6cc-691c906ee0d9)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_marketplace-operator-79b997595-4cnjs_openshift-marketplace_aec64a2e-ff83-4d35-a6cc-691c906ee0d9_0(ef18c54c47c7b3c34a22a9810c2da42d79dc33698d658fb0fe0e34254be408ac): error adding pod openshift-marketplace_marketplace-operator-79b997595-4cnjs to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"ef18c54c47c7b3c34a22a9810c2da42d79dc33698d658fb0fe0e34254be408ac\\\" Netns:\\\"/var/run/netns/a5a51a98-9d23-4650-bbb4-f21904646377\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-marketplace;K8S_POD_NAME=marketplace-operator-79b997595-4cnjs;K8S_POD_INFRA_CONTAINER_ID=ef18c54c47c7b3c34a22a9810c2da42d79dc33698d658fb0fe0e34254be408ac;K8S_POD_UID=aec64a2e-ff83-4d35-a6cc-691c906ee0d9\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-marketplace/marketplace-operator-79b997595-4cnjs] networking: Multus: [openshift-marketplace/marketplace-operator-79b997595-4cnjs/aec64a2e-ff83-4d35-a6cc-691c906ee0d9]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: SetNetworkStatus: failed to update the pod marketplace-operator-79b997595-4cnjs in out of cluster comm: status update failed for pod /: Get \\\"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-4cnjs?timeout=1m0s\\\": dial tcp 38.102.83.180:6443: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" podUID="aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.273904 4857 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="49bbb8a6b1c5f4cdbf237c00377c0f33bc21e2c5ce667991d691c0e1931f8af1" exitCode=0 Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.274004 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"49bbb8a6b1c5f4cdbf237c00377c0f33bc21e2c5ce667991d691c0e1931f8af1"} Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.274164 4857 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8bfed332-8a77-48dd-b3c6-7068b918dc71" Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.274183 4857 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8bfed332-8a77-48dd-b3c6-7068b918dc71" Feb 22 00:10:45 crc kubenswrapper[4857]: E0222 00:10:45.274517 4857 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.274519 4857 status_manager.go:851] "Failed to get status for pod" podUID="fa45663f-d64c-431c-be1d-29175cc804ea" pod="openshift-marketplace/community-operators-5rpdr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-5rpdr\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.274914 4857 status_manager.go:851] "Failed to get status for pod" podUID="10f2d337-12f1-4949-9872-08df9aa084ab" pod="openshift-marketplace/redhat-marketplace-797dx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-797dx\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.275396 4857 status_manager.go:851] "Failed to get status for pod" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" pod="openshift-marketplace/redhat-operators-xd4g7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xd4g7\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.275626 4857 status_manager.go:851] "Failed to get status for pod" podUID="18611d84-22b3-49df-9569-ac383eb42cab" pod="openshift-marketplace/certified-operators-m5rft" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-m5rft\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.275972 4857 status_manager.go:851] "Failed to get status for pod" podUID="4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-vmk2g\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.276422 4857 status_manager.go:851] "Failed to get status for pod" podUID="7ab5d710-d809-4d86-b56b-1f611cb02644" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.277680 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.277720 4857 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11" exitCode=1 Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.277748 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11"} Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.278089 4857 scope.go:117] "RemoveContainer" containerID="f879031b793d81fcf759f692048c698a6ace888736c08504b0ee6279bf904c11" Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.278377 4857 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.279264 4857 status_manager.go:851] "Failed to get status for pod" podUID="4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" pod="openshift-marketplace/marketplace-operator-79b997595-vmk2g" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/marketplace-operator-79b997595-vmk2g\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.279578 4857 status_manager.go:851] "Failed to get status for pod" podUID="7ab5d710-d809-4d86-b56b-1f611cb02644" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.279851 4857 status_manager.go:851] "Failed to get status for pod" podUID="fa45663f-d64c-431c-be1d-29175cc804ea" pod="openshift-marketplace/community-operators-5rpdr" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-5rpdr\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.280062 4857 status_manager.go:851] "Failed to get status for pod" podUID="10f2d337-12f1-4949-9872-08df9aa084ab" pod="openshift-marketplace/redhat-marketplace-797dx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-797dx\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.280252 4857 status_manager.go:851] "Failed to get status for pod" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" pod="openshift-marketplace/redhat-operators-xd4g7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xd4g7\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:45 crc kubenswrapper[4857]: I0222 00:10:45.280612 4857 status_manager.go:851] "Failed to get status for pod" podUID="18611d84-22b3-49df-9569-ac383eb42cab" pod="openshift-marketplace/certified-operators-m5rft" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-m5rft\": dial tcp 38.102.83.180:6443: connect: connection refused" Feb 22 00:10:46 crc kubenswrapper[4857]: I0222 00:10:46.290269 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5d7f5f87925b740e5110cec8bab28c20dc67808dcb38ed9931ab591c989631a9"} Feb 22 00:10:46 crc kubenswrapper[4857]: I0222 00:10:46.290584 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a1e69f61cbb530b35e85a6ab68b47dfe28828d3e184918b5f5af2d3c7315a3a0"} Feb 22 00:10:46 crc kubenswrapper[4857]: I0222 00:10:46.290596 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2826fc0b80b8b8a122c064b3d8efb2c6eabbd4834223883c44278ab28126323e"} Feb 22 00:10:46 crc kubenswrapper[4857]: I0222 00:10:46.290608 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4bef004ad05a9fe0b9eaf131db1354412ad572a2cd37cb105d197f893884887d"} Feb 22 00:10:46 crc kubenswrapper[4857]: I0222 00:10:46.293615 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 22 00:10:46 crc kubenswrapper[4857]: I0222 00:10:46.293654 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fa33136334f9e5d72197f9971dd3ae11936ca7ad19b6f2a5eb6a2e1442d0f5e3"} Feb 22 00:10:46 crc kubenswrapper[4857]: I0222 00:10:46.595422 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:10:46 crc kubenswrapper[4857]: I0222 00:10:46.604090 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:10:47 crc kubenswrapper[4857]: I0222 00:10:47.308822 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"11b725e85a393ef7489e8c75e194c10dfaeb9c14f70a9d721a95cb6ba39d538d"} Feb 22 00:10:47 crc kubenswrapper[4857]: I0222 00:10:47.308953 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:10:47 crc kubenswrapper[4857]: I0222 00:10:47.309060 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:47 crc kubenswrapper[4857]: I0222 00:10:47.309202 4857 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8bfed332-8a77-48dd-b3c6-7068b918dc71" Feb 22 00:10:47 crc kubenswrapper[4857]: I0222 00:10:47.309222 4857 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8bfed332-8a77-48dd-b3c6-7068b918dc71" Feb 22 00:10:49 crc kubenswrapper[4857]: I0222 00:10:49.160928 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:49 crc kubenswrapper[4857]: I0222 00:10:49.161081 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:49 crc kubenswrapper[4857]: I0222 00:10:49.166720 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:52 crc kubenswrapper[4857]: I0222 00:10:52.321293 4857 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:10:53 crc kubenswrapper[4857]: I0222 00:10:53.099996 4857 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="654c7563-4c05-4c83-ae9a-e41f58e31683" Feb 22 00:10:53 crc kubenswrapper[4857]: I0222 00:10:53.346989 4857 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8bfed332-8a77-48dd-b3c6-7068b918dc71" Feb 22 00:10:53 crc kubenswrapper[4857]: I0222 00:10:53.347090 4857 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8bfed332-8a77-48dd-b3c6-7068b918dc71" Feb 22 00:10:53 crc kubenswrapper[4857]: I0222 00:10:53.349598 4857 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="654c7563-4c05-4c83-ae9a-e41f58e31683" Feb 22 00:10:59 crc kubenswrapper[4857]: I0222 00:10:59.076997 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:10:59 crc kubenswrapper[4857]: I0222 00:10:59.078257 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:11:00 crc kubenswrapper[4857]: I0222 00:11:00.086056 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 22 00:11:00 crc kubenswrapper[4857]: I0222 00:11:00.389236 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4cnjs_aec64a2e-ff83-4d35-a6cc-691c906ee0d9/marketplace-operator/0.log" Feb 22 00:11:00 crc kubenswrapper[4857]: I0222 00:11:00.389302 4857 generic.go:334] "Generic (PLEG): container finished" podID="aec64a2e-ff83-4d35-a6cc-691c906ee0d9" containerID="899c148c46ea1d29aa82ba4edfefc3274652e25bd3cf17773e6864ed8a2c7673" exitCode=1 Feb 22 00:11:00 crc kubenswrapper[4857]: I0222 00:11:00.389345 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" event={"ID":"aec64a2e-ff83-4d35-a6cc-691c906ee0d9","Type":"ContainerDied","Data":"899c148c46ea1d29aa82ba4edfefc3274652e25bd3cf17773e6864ed8a2c7673"} Feb 22 00:11:00 crc kubenswrapper[4857]: I0222 00:11:00.389381 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" event={"ID":"aec64a2e-ff83-4d35-a6cc-691c906ee0d9","Type":"ContainerStarted","Data":"076f8835e390c263126903afb642397474104a0880004a2c859ae0d08aef1a57"} Feb 22 00:11:00 crc kubenswrapper[4857]: I0222 00:11:00.391000 4857 scope.go:117] "RemoveContainer" containerID="899c148c46ea1d29aa82ba4edfefc3274652e25bd3cf17773e6864ed8a2c7673" Feb 22 00:11:01 crc kubenswrapper[4857]: I0222 00:11:01.395871 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4cnjs_aec64a2e-ff83-4d35-a6cc-691c906ee0d9/marketplace-operator/1.log" Feb 22 00:11:01 crc kubenswrapper[4857]: I0222 00:11:01.396828 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4cnjs_aec64a2e-ff83-4d35-a6cc-691c906ee0d9/marketplace-operator/0.log" Feb 22 00:11:01 crc kubenswrapper[4857]: I0222 00:11:01.396885 4857 generic.go:334] "Generic (PLEG): container finished" podID="aec64a2e-ff83-4d35-a6cc-691c906ee0d9" containerID="082eed6b24c22f264d5ea4ce9c665a699df82cabd62789382aa3440b6fb4a3c9" exitCode=1 Feb 22 00:11:01 crc kubenswrapper[4857]: I0222 00:11:01.396924 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" event={"ID":"aec64a2e-ff83-4d35-a6cc-691c906ee0d9","Type":"ContainerDied","Data":"082eed6b24c22f264d5ea4ce9c665a699df82cabd62789382aa3440b6fb4a3c9"} Feb 22 00:11:01 crc kubenswrapper[4857]: I0222 00:11:01.396999 4857 scope.go:117] "RemoveContainer" containerID="899c148c46ea1d29aa82ba4edfefc3274652e25bd3cf17773e6864ed8a2c7673" Feb 22 00:11:01 crc kubenswrapper[4857]: I0222 00:11:01.398227 4857 scope.go:117] "RemoveContainer" containerID="082eed6b24c22f264d5ea4ce9c665a699df82cabd62789382aa3440b6fb4a3c9" Feb 22 00:11:01 crc kubenswrapper[4857]: E0222 00:11:01.399191 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-4cnjs_openshift-marketplace(aec64a2e-ff83-4d35-a6cc-691c906ee0d9)\"" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" podUID="aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Feb 22 00:11:01 crc kubenswrapper[4857]: I0222 00:11:01.523430 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 22 00:11:01 crc kubenswrapper[4857]: I0222 00:11:01.726674 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 22 00:11:02 crc kubenswrapper[4857]: I0222 00:11:02.005174 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 22 00:11:02 crc kubenswrapper[4857]: I0222 00:11:02.404303 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4cnjs_aec64a2e-ff83-4d35-a6cc-691c906ee0d9/marketplace-operator/1.log" Feb 22 00:11:02 crc kubenswrapper[4857]: I0222 00:11:02.731101 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 22 00:11:02 crc kubenswrapper[4857]: I0222 00:11:02.819001 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 22 00:11:02 crc kubenswrapper[4857]: I0222 00:11:02.946277 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 22 00:11:03 crc kubenswrapper[4857]: I0222 00:11:03.023707 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 22 00:11:03 crc kubenswrapper[4857]: I0222 00:11:03.046320 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 22 00:11:03 crc kubenswrapper[4857]: I0222 00:11:03.259981 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 22 00:11:03 crc kubenswrapper[4857]: I0222 00:11:03.331011 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 22 00:11:03 crc kubenswrapper[4857]: I0222 00:11:03.579932 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 22 00:11:03 crc kubenswrapper[4857]: I0222 00:11:03.648270 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 22 00:11:03 crc kubenswrapper[4857]: I0222 00:11:03.653003 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 22 00:11:03 crc kubenswrapper[4857]: I0222 00:11:03.750480 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 22 00:11:03 crc kubenswrapper[4857]: I0222 00:11:03.843007 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 22 00:11:03 crc kubenswrapper[4857]: I0222 00:11:03.856539 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 22 00:11:03 crc kubenswrapper[4857]: I0222 00:11:03.877217 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 22 00:11:03 crc kubenswrapper[4857]: I0222 00:11:03.916414 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 22 00:11:03 crc kubenswrapper[4857]: I0222 00:11:03.994424 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 22 00:11:04 crc kubenswrapper[4857]: I0222 00:11:04.354574 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 22 00:11:04 crc kubenswrapper[4857]: I0222 00:11:04.382651 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 22 00:11:04 crc kubenswrapper[4857]: I0222 00:11:04.382986 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 22 00:11:04 crc kubenswrapper[4857]: I0222 00:11:04.496469 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 22 00:11:04 crc kubenswrapper[4857]: I0222 00:11:04.669548 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 22 00:11:04 crc kubenswrapper[4857]: I0222 00:11:04.736494 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 22 00:11:04 crc kubenswrapper[4857]: I0222 00:11:04.780224 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 22 00:11:04 crc kubenswrapper[4857]: I0222 00:11:04.818855 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 22 00:11:04 crc kubenswrapper[4857]: I0222 00:11:04.860506 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 22 00:11:05 crc kubenswrapper[4857]: I0222 00:11:05.146505 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 22 00:11:05 crc kubenswrapper[4857]: I0222 00:11:05.214713 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 22 00:11:05 crc kubenswrapper[4857]: I0222 00:11:05.216733 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 22 00:11:05 crc kubenswrapper[4857]: I0222 00:11:05.232346 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 22 00:11:05 crc kubenswrapper[4857]: I0222 00:11:05.290424 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 22 00:11:05 crc kubenswrapper[4857]: I0222 00:11:05.433687 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 22 00:11:05 crc kubenswrapper[4857]: I0222 00:11:05.441356 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 22 00:11:05 crc kubenswrapper[4857]: I0222 00:11:05.509175 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 22 00:11:05 crc kubenswrapper[4857]: I0222 00:11:05.639418 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 22 00:11:05 crc kubenswrapper[4857]: I0222 00:11:05.777387 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 22 00:11:05 crc kubenswrapper[4857]: I0222 00:11:05.906514 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 22 00:11:05 crc kubenswrapper[4857]: I0222 00:11:05.971944 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 22 00:11:06 crc kubenswrapper[4857]: I0222 00:11:06.012133 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 22 00:11:06 crc kubenswrapper[4857]: I0222 00:11:06.101674 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 22 00:11:06 crc kubenswrapper[4857]: I0222 00:11:06.179164 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 22 00:11:06 crc kubenswrapper[4857]: I0222 00:11:06.190007 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 22 00:11:06 crc kubenswrapper[4857]: I0222 00:11:06.238812 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 22 00:11:06 crc kubenswrapper[4857]: I0222 00:11:06.327402 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 22 00:11:06 crc kubenswrapper[4857]: I0222 00:11:06.354618 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 22 00:11:06 crc kubenswrapper[4857]: I0222 00:11:06.413656 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 22 00:11:06 crc kubenswrapper[4857]: I0222 00:11:06.504609 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 22 00:11:06 crc kubenswrapper[4857]: I0222 00:11:06.690980 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 22 00:11:06 crc kubenswrapper[4857]: I0222 00:11:06.749610 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 22 00:11:06 crc kubenswrapper[4857]: I0222 00:11:06.793331 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 22 00:11:06 crc kubenswrapper[4857]: I0222 00:11:06.864624 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 22 00:11:06 crc kubenswrapper[4857]: I0222 00:11:06.866945 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 22 00:11:06 crc kubenswrapper[4857]: I0222 00:11:06.895566 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 22 00:11:06 crc kubenswrapper[4857]: I0222 00:11:06.934957 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 22 00:11:06 crc kubenswrapper[4857]: I0222 00:11:06.999380 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 22 00:11:07 crc kubenswrapper[4857]: I0222 00:11:07.036640 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 22 00:11:07 crc kubenswrapper[4857]: I0222 00:11:07.130626 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 22 00:11:07 crc kubenswrapper[4857]: I0222 00:11:07.234901 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 22 00:11:07 crc kubenswrapper[4857]: I0222 00:11:07.238864 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 22 00:11:07 crc kubenswrapper[4857]: I0222 00:11:07.319190 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 22 00:11:07 crc kubenswrapper[4857]: I0222 00:11:07.401581 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 22 00:11:07 crc kubenswrapper[4857]: I0222 00:11:07.417252 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 22 00:11:07 crc kubenswrapper[4857]: I0222 00:11:07.478765 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 22 00:11:07 crc kubenswrapper[4857]: I0222 00:11:07.498302 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 22 00:11:07 crc kubenswrapper[4857]: I0222 00:11:07.526107 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 22 00:11:07 crc kubenswrapper[4857]: I0222 00:11:07.530898 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 22 00:11:07 crc kubenswrapper[4857]: I0222 00:11:07.670293 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 22 00:11:07 crc kubenswrapper[4857]: I0222 00:11:07.676503 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 22 00:11:07 crc kubenswrapper[4857]: I0222 00:11:07.740853 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 22 00:11:07 crc kubenswrapper[4857]: I0222 00:11:07.768626 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 22 00:11:07 crc kubenswrapper[4857]: I0222 00:11:07.790608 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 22 00:11:08 crc kubenswrapper[4857]: I0222 00:11:08.039236 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 22 00:11:08 crc kubenswrapper[4857]: I0222 00:11:08.058198 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 22 00:11:08 crc kubenswrapper[4857]: I0222 00:11:08.058577 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 22 00:11:08 crc kubenswrapper[4857]: I0222 00:11:08.069849 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 22 00:11:08 crc kubenswrapper[4857]: I0222 00:11:08.226960 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 22 00:11:08 crc kubenswrapper[4857]: I0222 00:11:08.360790 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 22 00:11:08 crc kubenswrapper[4857]: I0222 00:11:08.361754 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 22 00:11:08 crc kubenswrapper[4857]: I0222 00:11:08.371179 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 22 00:11:08 crc kubenswrapper[4857]: I0222 00:11:08.440576 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 22 00:11:08 crc kubenswrapper[4857]: I0222 00:11:08.488755 4857 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 22 00:11:08 crc kubenswrapper[4857]: I0222 00:11:08.645910 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 22 00:11:08 crc kubenswrapper[4857]: I0222 00:11:08.729005 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.052410 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.103988 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.152152 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.242720 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.249538 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.266486 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.306325 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.315275 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.332109 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.352386 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.408628 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.413895 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.471641 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.492465 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.628072 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.841428 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.906930 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.914361 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.983508 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.984102 4857 scope.go:117] "RemoveContainer" containerID="082eed6b24c22f264d5ea4ce9c665a699df82cabd62789382aa3440b6fb4a3c9" Feb 22 00:11:09 crc kubenswrapper[4857]: I0222 00:11:09.984128 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:11:09 crc kubenswrapper[4857]: E0222 00:11:09.984403 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-4cnjs_openshift-marketplace(aec64a2e-ff83-4d35-a6cc-691c906ee0d9)\"" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" podUID="aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Feb 22 00:11:10 crc kubenswrapper[4857]: I0222 00:11:10.003104 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 22 00:11:10 crc kubenswrapper[4857]: I0222 00:11:10.038118 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 22 00:11:10 crc kubenswrapper[4857]: I0222 00:11:10.116890 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 22 00:11:10 crc kubenswrapper[4857]: I0222 00:11:10.186235 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 22 00:11:10 crc kubenswrapper[4857]: I0222 00:11:10.187301 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 22 00:11:10 crc kubenswrapper[4857]: I0222 00:11:10.193469 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 22 00:11:10 crc kubenswrapper[4857]: I0222 00:11:10.194815 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 22 00:11:10 crc kubenswrapper[4857]: I0222 00:11:10.301434 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 22 00:11:10 crc kubenswrapper[4857]: I0222 00:11:10.447021 4857 scope.go:117] "RemoveContainer" containerID="082eed6b24c22f264d5ea4ce9c665a699df82cabd62789382aa3440b6fb4a3c9" Feb 22 00:11:10 crc kubenswrapper[4857]: E0222 00:11:10.447320 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-4cnjs_openshift-marketplace(aec64a2e-ff83-4d35-a6cc-691c906ee0d9)\"" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" podUID="aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Feb 22 00:11:10 crc kubenswrapper[4857]: I0222 00:11:10.485901 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 22 00:11:10 crc kubenswrapper[4857]: I0222 00:11:10.506887 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 22 00:11:10 crc kubenswrapper[4857]: I0222 00:11:10.562986 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 22 00:11:10 crc kubenswrapper[4857]: I0222 00:11:10.600852 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 22 00:11:10 crc kubenswrapper[4857]: I0222 00:11:10.609975 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 22 00:11:10 crc kubenswrapper[4857]: I0222 00:11:10.754586 4857 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 22 00:11:10 crc kubenswrapper[4857]: I0222 00:11:10.808677 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 22 00:11:10 crc kubenswrapper[4857]: I0222 00:11:10.848112 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.006104 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.164233 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.169609 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.179019 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.245667 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.253270 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.361504 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.382759 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.429613 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.470373 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.620909 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.670633 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.693637 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.725194 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.765925 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.775799 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.785784 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.812596 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.865532 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.888123 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.918881 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.936486 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 22 00:11:11 crc kubenswrapper[4857]: I0222 00:11:11.978659 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.231356 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.251944 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.345568 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.352847 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.378058 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.394173 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.454999 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.468493 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.474467 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.519988 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.536342 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.647426 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.710207 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.733801 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.741266 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.764748 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.830004 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.854633 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 22 00:11:12 crc kubenswrapper[4857]: I0222 00:11:12.949957 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:12.996199 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.049689 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.096374 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.124920 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.153013 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.182084 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.298344 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.323962 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.328161 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.338019 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.344532 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.365365 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.400591 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.459947 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.507028 4857 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.511648 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vmk2g","openshift-marketplace/redhat-marketplace-797dx","openshift-kube-apiserver/kube-apiserver-crc","openshift-marketplace/redhat-operators-xd4g7","openshift-marketplace/community-operators-5rpdr","openshift-marketplace/certified-operators-m5rft"] Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.511712 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.511729 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4cnjs"] Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.512195 4857 scope.go:117] "RemoveContainer" containerID="082eed6b24c22f264d5ea4ce9c665a699df82cabd62789382aa3440b6fb4a3c9" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.517819 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.519341 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.545489 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.545469463 podStartE2EDuration="21.545469463s" podCreationTimestamp="2026-02-22 00:10:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:11:13.530095322 +0000 UTC m=+281.168824645" watchObservedRunningTime="2026-02-22 00:11:13.545469463 +0000 UTC m=+281.184198716" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.555263 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.620423 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.640608 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.656443 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.667074 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.687019 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.766917 4857 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 22 00:11:13 crc kubenswrapper[4857]: I0222 00:11:13.894865 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.005666 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.035465 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.088957 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.098937 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.142728 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.149092 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.198735 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.305115 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.324408 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.349844 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.416012 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.428512 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.469701 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4cnjs_aec64a2e-ff83-4d35-a6cc-691c906ee0d9/marketplace-operator/2.log" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.470178 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4cnjs_aec64a2e-ff83-4d35-a6cc-691c906ee0d9/marketplace-operator/1.log" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.470231 4857 generic.go:334] "Generic (PLEG): container finished" podID="aec64a2e-ff83-4d35-a6cc-691c906ee0d9" containerID="6c32e14cea229001079eb947f46c65893ce9c32f0ed10fbf37efd986a5e34c47" exitCode=1 Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.470315 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" event={"ID":"aec64a2e-ff83-4d35-a6cc-691c906ee0d9","Type":"ContainerDied","Data":"6c32e14cea229001079eb947f46c65893ce9c32f0ed10fbf37efd986a5e34c47"} Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.470365 4857 scope.go:117] "RemoveContainer" containerID="082eed6b24c22f264d5ea4ce9c665a699df82cabd62789382aa3440b6fb4a3c9" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.470959 4857 scope.go:117] "RemoveContainer" containerID="6c32e14cea229001079eb947f46c65893ce9c32f0ed10fbf37efd986a5e34c47" Feb 22 00:11:14 crc kubenswrapper[4857]: E0222 00:11:14.471245 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-4cnjs_openshift-marketplace(aec64a2e-ff83-4d35-a6cc-691c906ee0d9)\"" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" podUID="aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.566339 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.612356 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.672353 4857 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.672609 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://6f772d5bf4afd3597656f75b78ddd28f1cc5111bfbe4e0056153e48b09206633" gracePeriod=5 Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.727794 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.852245 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.868380 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.935541 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 22 00:11:14 crc kubenswrapper[4857]: I0222 00:11:14.979398 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.060231 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.083445 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10f2d337-12f1-4949-9872-08df9aa084ab" path="/var/lib/kubelet/pods/10f2d337-12f1-4949-9872-08df9aa084ab/volumes" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.084227 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18611d84-22b3-49df-9569-ac383eb42cab" path="/var/lib/kubelet/pods/18611d84-22b3-49df-9569-ac383eb42cab/volumes" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.084803 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.084874 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8" path="/var/lib/kubelet/pods/4511cf26-2444-4c3a-b5f8-4f9fd76ffdf8/volumes" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.085874 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" path="/var/lib/kubelet/pods/6d73807b-5e0c-4607-8393-5123a6cdbf9b/volumes" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.086544 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa45663f-d64c-431c-be1d-29175cc804ea" path="/var/lib/kubelet/pods/fa45663f-d64c-431c-be1d-29175cc804ea/volumes" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.197745 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.271185 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.299252 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.303998 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.332229 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.343948 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.373432 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.448255 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.482222 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.494334 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4cnjs_aec64a2e-ff83-4d35-a6cc-691c906ee0d9/marketplace-operator/2.log" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.628633 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.702181 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.805585 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.815911 4857 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.887019 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.893874 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 22 00:11:15 crc kubenswrapper[4857]: I0222 00:11:15.977560 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 22 00:11:16 crc kubenswrapper[4857]: I0222 00:11:16.008584 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 22 00:11:16 crc kubenswrapper[4857]: I0222 00:11:16.106408 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 22 00:11:16 crc kubenswrapper[4857]: I0222 00:11:16.108165 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 22 00:11:16 crc kubenswrapper[4857]: I0222 00:11:16.120980 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 22 00:11:16 crc kubenswrapper[4857]: I0222 00:11:16.202319 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 22 00:11:16 crc kubenswrapper[4857]: I0222 00:11:16.263546 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 22 00:11:16 crc kubenswrapper[4857]: I0222 00:11:16.392329 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 22 00:11:16 crc kubenswrapper[4857]: I0222 00:11:16.434571 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 22 00:11:16 crc kubenswrapper[4857]: I0222 00:11:16.469957 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 22 00:11:16 crc kubenswrapper[4857]: I0222 00:11:16.657752 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 22 00:11:16 crc kubenswrapper[4857]: I0222 00:11:16.689365 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 22 00:11:16 crc kubenswrapper[4857]: I0222 00:11:16.732864 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 22 00:11:17 crc kubenswrapper[4857]: I0222 00:11:17.033668 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 22 00:11:17 crc kubenswrapper[4857]: I0222 00:11:17.053961 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 22 00:11:17 crc kubenswrapper[4857]: I0222 00:11:17.126192 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 22 00:11:17 crc kubenswrapper[4857]: I0222 00:11:17.143880 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 22 00:11:17 crc kubenswrapper[4857]: I0222 00:11:17.542756 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 22 00:11:17 crc kubenswrapper[4857]: I0222 00:11:17.598265 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 22 00:11:17 crc kubenswrapper[4857]: I0222 00:11:17.697914 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 22 00:11:17 crc kubenswrapper[4857]: I0222 00:11:17.779382 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 22 00:11:18 crc kubenswrapper[4857]: I0222 00:11:18.124537 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 22 00:11:18 crc kubenswrapper[4857]: I0222 00:11:18.557181 4857 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 22 00:11:18 crc kubenswrapper[4857]: I0222 00:11:18.572265 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 22 00:11:19 crc kubenswrapper[4857]: I0222 00:11:19.983884 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:11:19 crc kubenswrapper[4857]: I0222 00:11:19.984759 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:11:19 crc kubenswrapper[4857]: I0222 00:11:19.985338 4857 scope.go:117] "RemoveContainer" containerID="6c32e14cea229001079eb947f46c65893ce9c32f0ed10fbf37efd986a5e34c47" Feb 22 00:11:19 crc kubenswrapper[4857]: E0222 00:11:19.985583 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-4cnjs_openshift-marketplace(aec64a2e-ff83-4d35-a6cc-691c906ee0d9)\"" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" podUID="aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Feb 22 00:11:20 crc kubenswrapper[4857]: I0222 00:11:20.520897 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 22 00:11:20 crc kubenswrapper[4857]: I0222 00:11:20.520949 4857 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="6f772d5bf4afd3597656f75b78ddd28f1cc5111bfbe4e0056153e48b09206633" exitCode=137 Feb 22 00:11:20 crc kubenswrapper[4857]: I0222 00:11:20.521076 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b37c92992e710d805b07ad9ffa78a5f9d7a8a756336f45750c9aee109e316b7" Feb 22 00:11:20 crc kubenswrapper[4857]: I0222 00:11:20.521468 4857 scope.go:117] "RemoveContainer" containerID="6c32e14cea229001079eb947f46c65893ce9c32f0ed10fbf37efd986a5e34c47" Feb 22 00:11:20 crc kubenswrapper[4857]: E0222 00:11:20.521674 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-4cnjs_openshift-marketplace(aec64a2e-ff83-4d35-a6cc-691c906ee0d9)\"" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" podUID="aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Feb 22 00:11:20 crc kubenswrapper[4857]: I0222 00:11:20.883010 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 22 00:11:20 crc kubenswrapper[4857]: I0222 00:11:20.883104 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:11:20 crc kubenswrapper[4857]: I0222 00:11:20.978448 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 22 00:11:20 crc kubenswrapper[4857]: I0222 00:11:20.978604 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:11:20 crc kubenswrapper[4857]: I0222 00:11:20.978785 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 22 00:11:20 crc kubenswrapper[4857]: I0222 00:11:20.978828 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 22 00:11:20 crc kubenswrapper[4857]: I0222 00:11:20.978838 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:11:20 crc kubenswrapper[4857]: I0222 00:11:20.978869 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 22 00:11:20 crc kubenswrapper[4857]: I0222 00:11:20.978904 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 22 00:11:20 crc kubenswrapper[4857]: I0222 00:11:20.979229 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:11:20 crc kubenswrapper[4857]: I0222 00:11:20.979305 4857 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 22 00:11:20 crc kubenswrapper[4857]: I0222 00:11:20.979376 4857 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 22 00:11:20 crc kubenswrapper[4857]: I0222 00:11:20.979269 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:11:20 crc kubenswrapper[4857]: I0222 00:11:20.988206 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:11:21 crc kubenswrapper[4857]: I0222 00:11:21.080298 4857 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 22 00:11:21 crc kubenswrapper[4857]: I0222 00:11:21.080347 4857 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 22 00:11:21 crc kubenswrapper[4857]: I0222 00:11:21.080358 4857 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 22 00:11:21 crc kubenswrapper[4857]: I0222 00:11:21.091407 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 22 00:11:21 crc kubenswrapper[4857]: I0222 00:11:21.524660 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 22 00:11:25 crc kubenswrapper[4857]: I0222 00:11:25.960637 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 22 00:11:28 crc kubenswrapper[4857]: I0222 00:11:28.936676 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 22 00:11:32 crc kubenswrapper[4857]: I0222 00:11:32.808627 4857 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Feb 22 00:11:33 crc kubenswrapper[4857]: I0222 00:11:33.080303 4857 scope.go:117] "RemoveContainer" containerID="6c32e14cea229001079eb947f46c65893ce9c32f0ed10fbf37efd986a5e34c47" Feb 22 00:11:33 crc kubenswrapper[4857]: E0222 00:11:33.080533 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"marketplace-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=marketplace-operator pod=marketplace-operator-79b997595-4cnjs_openshift-marketplace(aec64a2e-ff83-4d35-a6cc-691c906ee0d9)\"" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" podUID="aec64a2e-ff83-4d35-a6cc-691c906ee0d9" Feb 22 00:11:33 crc kubenswrapper[4857]: I0222 00:11:33.477533 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 22 00:11:35 crc kubenswrapper[4857]: I0222 00:11:35.714316 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 22 00:11:48 crc kubenswrapper[4857]: I0222 00:11:48.078209 4857 scope.go:117] "RemoveContainer" containerID="6c32e14cea229001079eb947f46c65893ce9c32f0ed10fbf37efd986a5e34c47" Feb 22 00:11:48 crc kubenswrapper[4857]: I0222 00:11:48.690126 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4cnjs_aec64a2e-ff83-4d35-a6cc-691c906ee0d9/marketplace-operator/2.log" Feb 22 00:11:48 crc kubenswrapper[4857]: I0222 00:11:48.690440 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" event={"ID":"aec64a2e-ff83-4d35-a6cc-691c906ee0d9","Type":"ContainerStarted","Data":"ebfa94f0665113db80274bf187509682db05a459a6fa8aebd6223261cc191411"} Feb 22 00:11:48 crc kubenswrapper[4857]: I0222 00:11:48.690707 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:11:48 crc kubenswrapper[4857]: I0222 00:11:48.695637 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" Feb 22 00:11:48 crc kubenswrapper[4857]: I0222 00:11:48.707149 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4cnjs" podStartSLOduration=79.707133128 podStartE2EDuration="1m19.707133128s" podCreationTimestamp="2026-02-22 00:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:11:48.706801388 +0000 UTC m=+316.345530641" watchObservedRunningTime="2026-02-22 00:11:48.707133128 +0000 UTC m=+316.345862391" Feb 22 00:11:50 crc kubenswrapper[4857]: I0222 00:11:50.533430 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-c669b"] Feb 22 00:11:50 crc kubenswrapper[4857]: I0222 00:11:50.533956 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" podUID="e8326fb9-1a3f-402a-acf2-62f7d6e82411" containerName="controller-manager" containerID="cri-o://b967b3799020d96b36d89771640a99805c2f6655ec137db205835b0199ff5a12" gracePeriod=30 Feb 22 00:11:50 crc kubenswrapper[4857]: I0222 00:11:50.653581 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v"] Feb 22 00:11:50 crc kubenswrapper[4857]: I0222 00:11:50.653816 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" podUID="60d12108-1ec8-4074-a60c-2b143a83f59f" containerName="route-controller-manager" containerID="cri-o://7fed6a816beba39e0dfd87ea70ce1f57f5b13bb463fd560b0dc98a55e8cde5b9" gracePeriod=30 Feb 22 00:11:50 crc kubenswrapper[4857]: I0222 00:11:50.701206 4857 generic.go:334] "Generic (PLEG): container finished" podID="e8326fb9-1a3f-402a-acf2-62f7d6e82411" containerID="b967b3799020d96b36d89771640a99805c2f6655ec137db205835b0199ff5a12" exitCode=0 Feb 22 00:11:50 crc kubenswrapper[4857]: I0222 00:11:50.701286 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" event={"ID":"e8326fb9-1a3f-402a-acf2-62f7d6e82411","Type":"ContainerDied","Data":"b967b3799020d96b36d89771640a99805c2f6655ec137db205835b0199ff5a12"} Feb 22 00:11:50 crc kubenswrapper[4857]: I0222 00:11:50.908230 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:11:50 crc kubenswrapper[4857]: I0222 00:11:50.959360 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.012596 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvbzl\" (UniqueName: \"kubernetes.io/projected/e8326fb9-1a3f-402a-acf2-62f7d6e82411-kube-api-access-jvbzl\") pod \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.012658 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60d12108-1ec8-4074-a60c-2b143a83f59f-config\") pod \"60d12108-1ec8-4074-a60c-2b143a83f59f\" (UID: \"60d12108-1ec8-4074-a60c-2b143a83f59f\") " Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.012699 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8326fb9-1a3f-402a-acf2-62f7d6e82411-serving-cert\") pod \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.012748 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60d12108-1ec8-4074-a60c-2b143a83f59f-client-ca\") pod \"60d12108-1ec8-4074-a60c-2b143a83f59f\" (UID: \"60d12108-1ec8-4074-a60c-2b143a83f59f\") " Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.012770 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tths\" (UniqueName: \"kubernetes.io/projected/60d12108-1ec8-4074-a60c-2b143a83f59f-kube-api-access-7tths\") pod \"60d12108-1ec8-4074-a60c-2b143a83f59f\" (UID: \"60d12108-1ec8-4074-a60c-2b143a83f59f\") " Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.012790 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60d12108-1ec8-4074-a60c-2b143a83f59f-serving-cert\") pod \"60d12108-1ec8-4074-a60c-2b143a83f59f\" (UID: \"60d12108-1ec8-4074-a60c-2b143a83f59f\") " Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.012810 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e8326fb9-1a3f-402a-acf2-62f7d6e82411-client-ca\") pod \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.012844 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8326fb9-1a3f-402a-acf2-62f7d6e82411-config\") pod \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.012868 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e8326fb9-1a3f-402a-acf2-62f7d6e82411-proxy-ca-bundles\") pod \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\" (UID: \"e8326fb9-1a3f-402a-acf2-62f7d6e82411\") " Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.013948 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8326fb9-1a3f-402a-acf2-62f7d6e82411-client-ca" (OuterVolumeSpecName: "client-ca") pod "e8326fb9-1a3f-402a-acf2-62f7d6e82411" (UID: "e8326fb9-1a3f-402a-acf2-62f7d6e82411"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.014188 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8326fb9-1a3f-402a-acf2-62f7d6e82411-config" (OuterVolumeSpecName: "config") pod "e8326fb9-1a3f-402a-acf2-62f7d6e82411" (UID: "e8326fb9-1a3f-402a-acf2-62f7d6e82411"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.014608 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60d12108-1ec8-4074-a60c-2b143a83f59f-client-ca" (OuterVolumeSpecName: "client-ca") pod "60d12108-1ec8-4074-a60c-2b143a83f59f" (UID: "60d12108-1ec8-4074-a60c-2b143a83f59f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.014749 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8326fb9-1a3f-402a-acf2-62f7d6e82411-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e8326fb9-1a3f-402a-acf2-62f7d6e82411" (UID: "e8326fb9-1a3f-402a-acf2-62f7d6e82411"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.014780 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60d12108-1ec8-4074-a60c-2b143a83f59f-config" (OuterVolumeSpecName: "config") pod "60d12108-1ec8-4074-a60c-2b143a83f59f" (UID: "60d12108-1ec8-4074-a60c-2b143a83f59f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.018839 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60d12108-1ec8-4074-a60c-2b143a83f59f-kube-api-access-7tths" (OuterVolumeSpecName: "kube-api-access-7tths") pod "60d12108-1ec8-4074-a60c-2b143a83f59f" (UID: "60d12108-1ec8-4074-a60c-2b143a83f59f"). InnerVolumeSpecName "kube-api-access-7tths". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.018908 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8326fb9-1a3f-402a-acf2-62f7d6e82411-kube-api-access-jvbzl" (OuterVolumeSpecName: "kube-api-access-jvbzl") pod "e8326fb9-1a3f-402a-acf2-62f7d6e82411" (UID: "e8326fb9-1a3f-402a-acf2-62f7d6e82411"). InnerVolumeSpecName "kube-api-access-jvbzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.018999 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8326fb9-1a3f-402a-acf2-62f7d6e82411-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e8326fb9-1a3f-402a-acf2-62f7d6e82411" (UID: "e8326fb9-1a3f-402a-acf2-62f7d6e82411"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.019052 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60d12108-1ec8-4074-a60c-2b143a83f59f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "60d12108-1ec8-4074-a60c-2b143a83f59f" (UID: "60d12108-1ec8-4074-a60c-2b143a83f59f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.114260 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvbzl\" (UniqueName: \"kubernetes.io/projected/e8326fb9-1a3f-402a-acf2-62f7d6e82411-kube-api-access-jvbzl\") on node \"crc\" DevicePath \"\"" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.114293 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60d12108-1ec8-4074-a60c-2b143a83f59f-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.114308 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8326fb9-1a3f-402a-acf2-62f7d6e82411-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.114320 4857 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60d12108-1ec8-4074-a60c-2b143a83f59f-client-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.114334 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tths\" (UniqueName: \"kubernetes.io/projected/60d12108-1ec8-4074-a60c-2b143a83f59f-kube-api-access-7tths\") on node \"crc\" DevicePath \"\"" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.114346 4857 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60d12108-1ec8-4074-a60c-2b143a83f59f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.114356 4857 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e8326fb9-1a3f-402a-acf2-62f7d6e82411-client-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.114366 4857 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8326fb9-1a3f-402a-acf2-62f7d6e82411-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.114379 4857 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e8326fb9-1a3f-402a-acf2-62f7d6e82411-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.708237 4857 generic.go:334] "Generic (PLEG): container finished" podID="60d12108-1ec8-4074-a60c-2b143a83f59f" containerID="7fed6a816beba39e0dfd87ea70ce1f57f5b13bb463fd560b0dc98a55e8cde5b9" exitCode=0 Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.708319 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" event={"ID":"60d12108-1ec8-4074-a60c-2b143a83f59f","Type":"ContainerDied","Data":"7fed6a816beba39e0dfd87ea70ce1f57f5b13bb463fd560b0dc98a55e8cde5b9"} Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.708716 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" event={"ID":"60d12108-1ec8-4074-a60c-2b143a83f59f","Type":"ContainerDied","Data":"49a036fb35334ac9f46de95f3f6cbedf36f22ab4de0f9a12eba85808c11618f0"} Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.708330 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.708740 4857 scope.go:117] "RemoveContainer" containerID="7fed6a816beba39e0dfd87ea70ce1f57f5b13bb463fd560b0dc98a55e8cde5b9" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.711311 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" event={"ID":"e8326fb9-1a3f-402a-acf2-62f7d6e82411","Type":"ContainerDied","Data":"056c2a52ec7cc59cccff869b4363924e9e2d3cf007b48b239c60317d275e2467"} Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.711400 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-c669b" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.730768 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v"] Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.747570 4857 scope.go:117] "RemoveContainer" containerID="7fed6a816beba39e0dfd87ea70ce1f57f5b13bb463fd560b0dc98a55e8cde5b9" Feb 22 00:11:51 crc kubenswrapper[4857]: E0222 00:11:51.749094 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fed6a816beba39e0dfd87ea70ce1f57f5b13bb463fd560b0dc98a55e8cde5b9\": container with ID starting with 7fed6a816beba39e0dfd87ea70ce1f57f5b13bb463fd560b0dc98a55e8cde5b9 not found: ID does not exist" containerID="7fed6a816beba39e0dfd87ea70ce1f57f5b13bb463fd560b0dc98a55e8cde5b9" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.749153 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fed6a816beba39e0dfd87ea70ce1f57f5b13bb463fd560b0dc98a55e8cde5b9"} err="failed to get container status \"7fed6a816beba39e0dfd87ea70ce1f57f5b13bb463fd560b0dc98a55e8cde5b9\": rpc error: code = NotFound desc = could not find container \"7fed6a816beba39e0dfd87ea70ce1f57f5b13bb463fd560b0dc98a55e8cde5b9\": container with ID starting with 7fed6a816beba39e0dfd87ea70ce1f57f5b13bb463fd560b0dc98a55e8cde5b9 not found: ID does not exist" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.749186 4857 scope.go:117] "RemoveContainer" containerID="b967b3799020d96b36d89771640a99805c2f6655ec137db205835b0199ff5a12" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.753531 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-p979v"] Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.765304 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-767d57b7cb-j6dbg"] Feb 22 00:11:51 crc kubenswrapper[4857]: E0222 00:11:51.765938 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f2d337-12f1-4949-9872-08df9aa084ab" containerName="registry-server" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.765955 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f2d337-12f1-4949-9872-08df9aa084ab" containerName="registry-server" Feb 22 00:11:51 crc kubenswrapper[4857]: E0222 00:11:51.765983 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ab5d710-d809-4d86-b56b-1f611cb02644" containerName="installer" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.765991 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ab5d710-d809-4d86-b56b-1f611cb02644" containerName="installer" Feb 22 00:11:51 crc kubenswrapper[4857]: E0222 00:11:51.766003 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" containerName="registry-server" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.766011 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" containerName="registry-server" Feb 22 00:11:51 crc kubenswrapper[4857]: E0222 00:11:51.766068 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8326fb9-1a3f-402a-acf2-62f7d6e82411" containerName="controller-manager" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.766079 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8326fb9-1a3f-402a-acf2-62f7d6e82411" containerName="controller-manager" Feb 22 00:11:51 crc kubenswrapper[4857]: E0222 00:11:51.766101 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60d12108-1ec8-4074-a60c-2b143a83f59f" containerName="route-controller-manager" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.766111 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="60d12108-1ec8-4074-a60c-2b143a83f59f" containerName="route-controller-manager" Feb 22 00:11:51 crc kubenswrapper[4857]: E0222 00:11:51.766127 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f2d337-12f1-4949-9872-08df9aa084ab" containerName="extract-utilities" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.766135 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f2d337-12f1-4949-9872-08df9aa084ab" containerName="extract-utilities" Feb 22 00:11:51 crc kubenswrapper[4857]: E0222 00:11:51.766146 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" containerName="extract-utilities" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.766154 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" containerName="extract-utilities" Feb 22 00:11:51 crc kubenswrapper[4857]: E0222 00:11:51.766174 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.766183 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 22 00:11:51 crc kubenswrapper[4857]: E0222 00:11:51.766198 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" containerName="extract-content" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.766206 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" containerName="extract-content" Feb 22 00:11:51 crc kubenswrapper[4857]: E0222 00:11:51.766224 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f2d337-12f1-4949-9872-08df9aa084ab" containerName="extract-content" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.766234 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f2d337-12f1-4949-9872-08df9aa084ab" containerName="extract-content" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.766780 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="10f2d337-12f1-4949-9872-08df9aa084ab" containerName="registry-server" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.766817 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.766833 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8326fb9-1a3f-402a-acf2-62f7d6e82411" containerName="controller-manager" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.766854 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ab5d710-d809-4d86-b56b-1f611cb02644" containerName="installer" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.766863 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d73807b-5e0c-4607-8393-5123a6cdbf9b" containerName="registry-server" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.766884 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="60d12108-1ec8-4074-a60c-2b143a83f59f" containerName="route-controller-manager" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.767673 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.776700 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6"] Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.778910 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.779629 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.779893 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.780593 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.781542 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.782135 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.782348 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.784064 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6"] Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.785332 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.785348 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.785564 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.785642 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.785736 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.785943 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.787814 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.791589 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-767d57b7cb-j6dbg"] Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.803448 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-c669b"] Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.810544 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-c669b"] Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.823631 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41599ec7-6031-4f85-a9da-bdd92fb68d25-client-ca\") pod \"route-controller-manager-5bf5f5dc6d-6s4m6\" (UID: \"41599ec7-6031-4f85-a9da-bdd92fb68d25\") " pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.823707 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6b52796-fcdf-4499-8401-49a8d59d383b-config\") pod \"controller-manager-767d57b7cb-j6dbg\" (UID: \"c6b52796-fcdf-4499-8401-49a8d59d383b\") " pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.823851 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5gxp\" (UniqueName: \"kubernetes.io/projected/41599ec7-6031-4f85-a9da-bdd92fb68d25-kube-api-access-z5gxp\") pod \"route-controller-manager-5bf5f5dc6d-6s4m6\" (UID: \"41599ec7-6031-4f85-a9da-bdd92fb68d25\") " pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.823983 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c6b52796-fcdf-4499-8401-49a8d59d383b-client-ca\") pod \"controller-manager-767d57b7cb-j6dbg\" (UID: \"c6b52796-fcdf-4499-8401-49a8d59d383b\") " pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.824105 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6b52796-fcdf-4499-8401-49a8d59d383b-serving-cert\") pod \"controller-manager-767d57b7cb-j6dbg\" (UID: \"c6b52796-fcdf-4499-8401-49a8d59d383b\") " pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.824144 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4qqn\" (UniqueName: \"kubernetes.io/projected/c6b52796-fcdf-4499-8401-49a8d59d383b-kube-api-access-h4qqn\") pod \"controller-manager-767d57b7cb-j6dbg\" (UID: \"c6b52796-fcdf-4499-8401-49a8d59d383b\") " pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.824184 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41599ec7-6031-4f85-a9da-bdd92fb68d25-config\") pod \"route-controller-manager-5bf5f5dc6d-6s4m6\" (UID: \"41599ec7-6031-4f85-a9da-bdd92fb68d25\") " pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.824208 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c6b52796-fcdf-4499-8401-49a8d59d383b-proxy-ca-bundles\") pod \"controller-manager-767d57b7cb-j6dbg\" (UID: \"c6b52796-fcdf-4499-8401-49a8d59d383b\") " pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.824282 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41599ec7-6031-4f85-a9da-bdd92fb68d25-serving-cert\") pod \"route-controller-manager-5bf5f5dc6d-6s4m6\" (UID: \"41599ec7-6031-4f85-a9da-bdd92fb68d25\") " pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.925847 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6b52796-fcdf-4499-8401-49a8d59d383b-config\") pod \"controller-manager-767d57b7cb-j6dbg\" (UID: \"c6b52796-fcdf-4499-8401-49a8d59d383b\") " pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.925929 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5gxp\" (UniqueName: \"kubernetes.io/projected/41599ec7-6031-4f85-a9da-bdd92fb68d25-kube-api-access-z5gxp\") pod \"route-controller-manager-5bf5f5dc6d-6s4m6\" (UID: \"41599ec7-6031-4f85-a9da-bdd92fb68d25\") " pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.925975 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c6b52796-fcdf-4499-8401-49a8d59d383b-client-ca\") pod \"controller-manager-767d57b7cb-j6dbg\" (UID: \"c6b52796-fcdf-4499-8401-49a8d59d383b\") " pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.926018 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6b52796-fcdf-4499-8401-49a8d59d383b-serving-cert\") pod \"controller-manager-767d57b7cb-j6dbg\" (UID: \"c6b52796-fcdf-4499-8401-49a8d59d383b\") " pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.926068 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4qqn\" (UniqueName: \"kubernetes.io/projected/c6b52796-fcdf-4499-8401-49a8d59d383b-kube-api-access-h4qqn\") pod \"controller-manager-767d57b7cb-j6dbg\" (UID: \"c6b52796-fcdf-4499-8401-49a8d59d383b\") " pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.926126 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41599ec7-6031-4f85-a9da-bdd92fb68d25-config\") pod \"route-controller-manager-5bf5f5dc6d-6s4m6\" (UID: \"41599ec7-6031-4f85-a9da-bdd92fb68d25\") " pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.926156 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c6b52796-fcdf-4499-8401-49a8d59d383b-proxy-ca-bundles\") pod \"controller-manager-767d57b7cb-j6dbg\" (UID: \"c6b52796-fcdf-4499-8401-49a8d59d383b\") " pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.926191 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41599ec7-6031-4f85-a9da-bdd92fb68d25-serving-cert\") pod \"route-controller-manager-5bf5f5dc6d-6s4m6\" (UID: \"41599ec7-6031-4f85-a9da-bdd92fb68d25\") " pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.926227 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41599ec7-6031-4f85-a9da-bdd92fb68d25-client-ca\") pod \"route-controller-manager-5bf5f5dc6d-6s4m6\" (UID: \"41599ec7-6031-4f85-a9da-bdd92fb68d25\") " pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.927564 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41599ec7-6031-4f85-a9da-bdd92fb68d25-client-ca\") pod \"route-controller-manager-5bf5f5dc6d-6s4m6\" (UID: \"41599ec7-6031-4f85-a9da-bdd92fb68d25\") " pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.927583 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c6b52796-fcdf-4499-8401-49a8d59d383b-client-ca\") pod \"controller-manager-767d57b7cb-j6dbg\" (UID: \"c6b52796-fcdf-4499-8401-49a8d59d383b\") " pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.927759 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c6b52796-fcdf-4499-8401-49a8d59d383b-proxy-ca-bundles\") pod \"controller-manager-767d57b7cb-j6dbg\" (UID: \"c6b52796-fcdf-4499-8401-49a8d59d383b\") " pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.928181 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41599ec7-6031-4f85-a9da-bdd92fb68d25-config\") pod \"route-controller-manager-5bf5f5dc6d-6s4m6\" (UID: \"41599ec7-6031-4f85-a9da-bdd92fb68d25\") " pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.928509 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6b52796-fcdf-4499-8401-49a8d59d383b-config\") pod \"controller-manager-767d57b7cb-j6dbg\" (UID: \"c6b52796-fcdf-4499-8401-49a8d59d383b\") " pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.932894 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41599ec7-6031-4f85-a9da-bdd92fb68d25-serving-cert\") pod \"route-controller-manager-5bf5f5dc6d-6s4m6\" (UID: \"41599ec7-6031-4f85-a9da-bdd92fb68d25\") " pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.934760 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6b52796-fcdf-4499-8401-49a8d59d383b-serving-cert\") pod \"controller-manager-767d57b7cb-j6dbg\" (UID: \"c6b52796-fcdf-4499-8401-49a8d59d383b\") " pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.948183 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4qqn\" (UniqueName: \"kubernetes.io/projected/c6b52796-fcdf-4499-8401-49a8d59d383b-kube-api-access-h4qqn\") pod \"controller-manager-767d57b7cb-j6dbg\" (UID: \"c6b52796-fcdf-4499-8401-49a8d59d383b\") " pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:51 crc kubenswrapper[4857]: I0222 00:11:51.949656 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5gxp\" (UniqueName: \"kubernetes.io/projected/41599ec7-6031-4f85-a9da-bdd92fb68d25-kube-api-access-z5gxp\") pod \"route-controller-manager-5bf5f5dc6d-6s4m6\" (UID: \"41599ec7-6031-4f85-a9da-bdd92fb68d25\") " pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" Feb 22 00:11:52 crc kubenswrapper[4857]: I0222 00:11:52.097858 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:52 crc kubenswrapper[4857]: I0222 00:11:52.105169 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" Feb 22 00:11:52 crc kubenswrapper[4857]: I0222 00:11:52.319141 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-767d57b7cb-j6dbg"] Feb 22 00:11:52 crc kubenswrapper[4857]: W0222 00:11:52.324284 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6b52796_fcdf_4499_8401_49a8d59d383b.slice/crio-132596b799cc50e544fb8b4f13f261c55a93b875dd498863a9f0b33b105d1c45 WatchSource:0}: Error finding container 132596b799cc50e544fb8b4f13f261c55a93b875dd498863a9f0b33b105d1c45: Status 404 returned error can't find the container with id 132596b799cc50e544fb8b4f13f261c55a93b875dd498863a9f0b33b105d1c45 Feb 22 00:11:52 crc kubenswrapper[4857]: I0222 00:11:52.355728 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6"] Feb 22 00:11:52 crc kubenswrapper[4857]: W0222 00:11:52.370290 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41599ec7_6031_4f85_a9da_bdd92fb68d25.slice/crio-a4e9dea693fa915de9890619a726819197cf171ede67b7d9ac61b4a354740586 WatchSource:0}: Error finding container a4e9dea693fa915de9890619a726819197cf171ede67b7d9ac61b4a354740586: Status 404 returned error can't find the container with id a4e9dea693fa915de9890619a726819197cf171ede67b7d9ac61b4a354740586 Feb 22 00:11:52 crc kubenswrapper[4857]: I0222 00:11:52.718433 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" event={"ID":"c6b52796-fcdf-4499-8401-49a8d59d383b","Type":"ContainerStarted","Data":"a68c3ae1b628117dd1f9172171a41acf208f16b1ef407ff260e7d3f02c9820ea"} Feb 22 00:11:52 crc kubenswrapper[4857]: I0222 00:11:52.718483 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" event={"ID":"c6b52796-fcdf-4499-8401-49a8d59d383b","Type":"ContainerStarted","Data":"132596b799cc50e544fb8b4f13f261c55a93b875dd498863a9f0b33b105d1c45"} Feb 22 00:11:52 crc kubenswrapper[4857]: I0222 00:11:52.718843 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:52 crc kubenswrapper[4857]: I0222 00:11:52.722221 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" event={"ID":"41599ec7-6031-4f85-a9da-bdd92fb68d25","Type":"ContainerStarted","Data":"e15a802cee9e5bdef15d9f4893f98576a0cd0cb2a54d4c60e5235fb167672bc8"} Feb 22 00:11:52 crc kubenswrapper[4857]: I0222 00:11:52.722266 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" event={"ID":"41599ec7-6031-4f85-a9da-bdd92fb68d25","Type":"ContainerStarted","Data":"a4e9dea693fa915de9890619a726819197cf171ede67b7d9ac61b4a354740586"} Feb 22 00:11:52 crc kubenswrapper[4857]: I0222 00:11:52.722284 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" Feb 22 00:11:52 crc kubenswrapper[4857]: I0222 00:11:52.728164 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" Feb 22 00:11:52 crc kubenswrapper[4857]: I0222 00:11:52.741684 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-767d57b7cb-j6dbg" podStartSLOduration=2.741666748 podStartE2EDuration="2.741666748s" podCreationTimestamp="2026-02-22 00:11:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:11:52.738278894 +0000 UTC m=+320.377008147" watchObservedRunningTime="2026-02-22 00:11:52.741666748 +0000 UTC m=+320.380396001" Feb 22 00:11:53 crc kubenswrapper[4857]: I0222 00:11:53.084334 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60d12108-1ec8-4074-a60c-2b143a83f59f" path="/var/lib/kubelet/pods/60d12108-1ec8-4074-a60c-2b143a83f59f/volumes" Feb 22 00:11:53 crc kubenswrapper[4857]: I0222 00:11:53.085095 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8326fb9-1a3f-402a-acf2-62f7d6e82411" path="/var/lib/kubelet/pods/e8326fb9-1a3f-402a-acf2-62f7d6e82411/volumes" Feb 22 00:11:53 crc kubenswrapper[4857]: I0222 00:11:53.415607 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" Feb 22 00:11:53 crc kubenswrapper[4857]: I0222 00:11:53.438179 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5bf5f5dc6d-6s4m6" podStartSLOduration=3.438155751 podStartE2EDuration="3.438155751s" podCreationTimestamp="2026-02-22 00:11:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:11:52.782418408 +0000 UTC m=+320.421147671" watchObservedRunningTime="2026-02-22 00:11:53.438155751 +0000 UTC m=+321.076885014" Feb 22 00:12:10 crc kubenswrapper[4857]: I0222 00:12:10.427994 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:12:10 crc kubenswrapper[4857]: I0222 00:12:10.428392 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:12:40 crc kubenswrapper[4857]: I0222 00:12:40.428416 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:12:40 crc kubenswrapper[4857]: I0222 00:12:40.431546 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.169936 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kd7ww"] Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.171352 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kd7ww" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.173454 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.179957 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd7ww"] Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.239329 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swfzp\" (UniqueName: \"kubernetes.io/projected/b9df334b-3beb-41e4-9ac7-cfaaa10ad923-kube-api-access-swfzp\") pod \"redhat-marketplace-kd7ww\" (UID: \"b9df334b-3beb-41e4-9ac7-cfaaa10ad923\") " pod="openshift-marketplace/redhat-marketplace-kd7ww" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.239366 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9df334b-3beb-41e4-9ac7-cfaaa10ad923-utilities\") pod \"redhat-marketplace-kd7ww\" (UID: \"b9df334b-3beb-41e4-9ac7-cfaaa10ad923\") " pod="openshift-marketplace/redhat-marketplace-kd7ww" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.239386 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9df334b-3beb-41e4-9ac7-cfaaa10ad923-catalog-content\") pod \"redhat-marketplace-kd7ww\" (UID: \"b9df334b-3beb-41e4-9ac7-cfaaa10ad923\") " pod="openshift-marketplace/redhat-marketplace-kd7ww" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.340876 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swfzp\" (UniqueName: \"kubernetes.io/projected/b9df334b-3beb-41e4-9ac7-cfaaa10ad923-kube-api-access-swfzp\") pod \"redhat-marketplace-kd7ww\" (UID: \"b9df334b-3beb-41e4-9ac7-cfaaa10ad923\") " pod="openshift-marketplace/redhat-marketplace-kd7ww" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.340915 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9df334b-3beb-41e4-9ac7-cfaaa10ad923-utilities\") pod \"redhat-marketplace-kd7ww\" (UID: \"b9df334b-3beb-41e4-9ac7-cfaaa10ad923\") " pod="openshift-marketplace/redhat-marketplace-kd7ww" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.340936 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9df334b-3beb-41e4-9ac7-cfaaa10ad923-catalog-content\") pod \"redhat-marketplace-kd7ww\" (UID: \"b9df334b-3beb-41e4-9ac7-cfaaa10ad923\") " pod="openshift-marketplace/redhat-marketplace-kd7ww" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.341399 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9df334b-3beb-41e4-9ac7-cfaaa10ad923-catalog-content\") pod \"redhat-marketplace-kd7ww\" (UID: \"b9df334b-3beb-41e4-9ac7-cfaaa10ad923\") " pod="openshift-marketplace/redhat-marketplace-kd7ww" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.341940 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9df334b-3beb-41e4-9ac7-cfaaa10ad923-utilities\") pod \"redhat-marketplace-kd7ww\" (UID: \"b9df334b-3beb-41e4-9ac7-cfaaa10ad923\") " pod="openshift-marketplace/redhat-marketplace-kd7ww" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.359151 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swfzp\" (UniqueName: \"kubernetes.io/projected/b9df334b-3beb-41e4-9ac7-cfaaa10ad923-kube-api-access-swfzp\") pod \"redhat-marketplace-kd7ww\" (UID: \"b9df334b-3beb-41e4-9ac7-cfaaa10ad923\") " pod="openshift-marketplace/redhat-marketplace-kd7ww" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.366169 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xrdpk"] Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.367352 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xrdpk" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.369981 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.375843 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xrdpk"] Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.442021 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2-catalog-content\") pod \"redhat-operators-xrdpk\" (UID: \"ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2\") " pod="openshift-marketplace/redhat-operators-xrdpk" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.442111 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2-utilities\") pod \"redhat-operators-xrdpk\" (UID: \"ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2\") " pod="openshift-marketplace/redhat-operators-xrdpk" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.442135 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpgnd\" (UniqueName: \"kubernetes.io/projected/ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2-kube-api-access-bpgnd\") pod \"redhat-operators-xrdpk\" (UID: \"ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2\") " pod="openshift-marketplace/redhat-operators-xrdpk" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.485968 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kd7ww" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.551842 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2-utilities\") pod \"redhat-operators-xrdpk\" (UID: \"ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2\") " pod="openshift-marketplace/redhat-operators-xrdpk" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.551914 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpgnd\" (UniqueName: \"kubernetes.io/projected/ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2-kube-api-access-bpgnd\") pod \"redhat-operators-xrdpk\" (UID: \"ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2\") " pod="openshift-marketplace/redhat-operators-xrdpk" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.551990 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2-catalog-content\") pod \"redhat-operators-xrdpk\" (UID: \"ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2\") " pod="openshift-marketplace/redhat-operators-xrdpk" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.552435 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2-utilities\") pod \"redhat-operators-xrdpk\" (UID: \"ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2\") " pod="openshift-marketplace/redhat-operators-xrdpk" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.552465 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2-catalog-content\") pod \"redhat-operators-xrdpk\" (UID: \"ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2\") " pod="openshift-marketplace/redhat-operators-xrdpk" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.572614 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpgnd\" (UniqueName: \"kubernetes.io/projected/ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2-kube-api-access-bpgnd\") pod \"redhat-operators-xrdpk\" (UID: \"ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2\") " pod="openshift-marketplace/redhat-operators-xrdpk" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.709614 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xrdpk" Feb 22 00:12:46 crc kubenswrapper[4857]: I0222 00:12:46.859184 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd7ww"] Feb 22 00:12:47 crc kubenswrapper[4857]: I0222 00:12:47.071058 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd7ww" event={"ID":"b9df334b-3beb-41e4-9ac7-cfaaa10ad923","Type":"ContainerStarted","Data":"97c1c99e002c7300d437014c4873e99bca6ec77fbec1972093fa5cf70c4ae7d0"} Feb 22 00:12:47 crc kubenswrapper[4857]: I0222 00:12:47.150713 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xrdpk"] Feb 22 00:12:47 crc kubenswrapper[4857]: W0222 00:12:47.157433 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca1ea5d7_d3fa_4aa9_ad7e_393011f9d5f2.slice/crio-a4df633578f019bd5b5de1515b6bcd9a2d25906ebb6d49639280ad0ddd70274f WatchSource:0}: Error finding container a4df633578f019bd5b5de1515b6bcd9a2d25906ebb6d49639280ad0ddd70274f: Status 404 returned error can't find the container with id a4df633578f019bd5b5de1515b6bcd9a2d25906ebb6d49639280ad0ddd70274f Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.078892 4857 generic.go:334] "Generic (PLEG): container finished" podID="ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2" containerID="629aa94e2b699cb1a619219ac7d095f55d6893dab7d5d5dce4284470a04277e1" exitCode=0 Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.078948 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xrdpk" event={"ID":"ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2","Type":"ContainerDied","Data":"629aa94e2b699cb1a619219ac7d095f55d6893dab7d5d5dce4284470a04277e1"} Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.078970 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xrdpk" event={"ID":"ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2","Type":"ContainerStarted","Data":"a4df633578f019bd5b5de1515b6bcd9a2d25906ebb6d49639280ad0ddd70274f"} Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.081117 4857 generic.go:334] "Generic (PLEG): container finished" podID="b9df334b-3beb-41e4-9ac7-cfaaa10ad923" containerID="8e06bf1bbd274d6e83ebe407c102dc09261e04dbf127cebb87acd2ed0c796d52" exitCode=0 Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.081184 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd7ww" event={"ID":"b9df334b-3beb-41e4-9ac7-cfaaa10ad923","Type":"ContainerDied","Data":"8e06bf1bbd274d6e83ebe407c102dc09261e04dbf127cebb87acd2ed0c796d52"} Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.666193 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fwkxt"] Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.669033 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fwkxt" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.670745 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.673215 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fwkxt"] Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.766850 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nx245"] Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.767986 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nx245" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.769609 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.779251 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nx245"] Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.784968 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc489b84-7090-4294-84f5-77edb82e6ca0-utilities\") pod \"community-operators-fwkxt\" (UID: \"dc489b84-7090-4294-84f5-77edb82e6ca0\") " pod="openshift-marketplace/community-operators-fwkxt" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.785074 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzs67\" (UniqueName: \"kubernetes.io/projected/dc489b84-7090-4294-84f5-77edb82e6ca0-kube-api-access-jzs67\") pod \"community-operators-fwkxt\" (UID: \"dc489b84-7090-4294-84f5-77edb82e6ca0\") " pod="openshift-marketplace/community-operators-fwkxt" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.785423 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc489b84-7090-4294-84f5-77edb82e6ca0-catalog-content\") pod \"community-operators-fwkxt\" (UID: \"dc489b84-7090-4294-84f5-77edb82e6ca0\") " pod="openshift-marketplace/community-operators-fwkxt" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.887700 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc489b84-7090-4294-84f5-77edb82e6ca0-utilities\") pod \"community-operators-fwkxt\" (UID: \"dc489b84-7090-4294-84f5-77edb82e6ca0\") " pod="openshift-marketplace/community-operators-fwkxt" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.887809 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzs67\" (UniqueName: \"kubernetes.io/projected/dc489b84-7090-4294-84f5-77edb82e6ca0-kube-api-access-jzs67\") pod \"community-operators-fwkxt\" (UID: \"dc489b84-7090-4294-84f5-77edb82e6ca0\") " pod="openshift-marketplace/community-operators-fwkxt" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.887868 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phnnz\" (UniqueName: \"kubernetes.io/projected/361dcb13-b990-403e-814b-cf0d11668bbf-kube-api-access-phnnz\") pod \"certified-operators-nx245\" (UID: \"361dcb13-b990-403e-814b-cf0d11668bbf\") " pod="openshift-marketplace/certified-operators-nx245" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.887906 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/361dcb13-b990-403e-814b-cf0d11668bbf-catalog-content\") pod \"certified-operators-nx245\" (UID: \"361dcb13-b990-403e-814b-cf0d11668bbf\") " pod="openshift-marketplace/certified-operators-nx245" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.887982 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc489b84-7090-4294-84f5-77edb82e6ca0-catalog-content\") pod \"community-operators-fwkxt\" (UID: \"dc489b84-7090-4294-84f5-77edb82e6ca0\") " pod="openshift-marketplace/community-operators-fwkxt" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.888120 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/361dcb13-b990-403e-814b-cf0d11668bbf-utilities\") pod \"certified-operators-nx245\" (UID: \"361dcb13-b990-403e-814b-cf0d11668bbf\") " pod="openshift-marketplace/certified-operators-nx245" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.888453 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc489b84-7090-4294-84f5-77edb82e6ca0-utilities\") pod \"community-operators-fwkxt\" (UID: \"dc489b84-7090-4294-84f5-77edb82e6ca0\") " pod="openshift-marketplace/community-operators-fwkxt" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.888515 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc489b84-7090-4294-84f5-77edb82e6ca0-catalog-content\") pod \"community-operators-fwkxt\" (UID: \"dc489b84-7090-4294-84f5-77edb82e6ca0\") " pod="openshift-marketplace/community-operators-fwkxt" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.917873 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzs67\" (UniqueName: \"kubernetes.io/projected/dc489b84-7090-4294-84f5-77edb82e6ca0-kube-api-access-jzs67\") pod \"community-operators-fwkxt\" (UID: \"dc489b84-7090-4294-84f5-77edb82e6ca0\") " pod="openshift-marketplace/community-operators-fwkxt" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.989280 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/361dcb13-b990-403e-814b-cf0d11668bbf-utilities\") pod \"certified-operators-nx245\" (UID: \"361dcb13-b990-403e-814b-cf0d11668bbf\") " pod="openshift-marketplace/certified-operators-nx245" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.989392 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phnnz\" (UniqueName: \"kubernetes.io/projected/361dcb13-b990-403e-814b-cf0d11668bbf-kube-api-access-phnnz\") pod \"certified-operators-nx245\" (UID: \"361dcb13-b990-403e-814b-cf0d11668bbf\") " pod="openshift-marketplace/certified-operators-nx245" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.989421 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/361dcb13-b990-403e-814b-cf0d11668bbf-catalog-content\") pod \"certified-operators-nx245\" (UID: \"361dcb13-b990-403e-814b-cf0d11668bbf\") " pod="openshift-marketplace/certified-operators-nx245" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.989990 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/361dcb13-b990-403e-814b-cf0d11668bbf-utilities\") pod \"certified-operators-nx245\" (UID: \"361dcb13-b990-403e-814b-cf0d11668bbf\") " pod="openshift-marketplace/certified-operators-nx245" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.990127 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/361dcb13-b990-403e-814b-cf0d11668bbf-catalog-content\") pod \"certified-operators-nx245\" (UID: \"361dcb13-b990-403e-814b-cf0d11668bbf\") " pod="openshift-marketplace/certified-operators-nx245" Feb 22 00:12:48 crc kubenswrapper[4857]: I0222 00:12:48.993887 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fwkxt" Feb 22 00:12:49 crc kubenswrapper[4857]: I0222 00:12:49.009154 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phnnz\" (UniqueName: \"kubernetes.io/projected/361dcb13-b990-403e-814b-cf0d11668bbf-kube-api-access-phnnz\") pod \"certified-operators-nx245\" (UID: \"361dcb13-b990-403e-814b-cf0d11668bbf\") " pod="openshift-marketplace/certified-operators-nx245" Feb 22 00:12:49 crc kubenswrapper[4857]: I0222 00:12:49.083590 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nx245" Feb 22 00:12:49 crc kubenswrapper[4857]: I0222 00:12:49.207224 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fwkxt"] Feb 22 00:12:49 crc kubenswrapper[4857]: W0222 00:12:49.213899 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc489b84_7090_4294_84f5_77edb82e6ca0.slice/crio-a542b212cf1a53f9b3d07614d5ee19349ca198ceeddee29d9e8276b05190e505 WatchSource:0}: Error finding container a542b212cf1a53f9b3d07614d5ee19349ca198ceeddee29d9e8276b05190e505: Status 404 returned error can't find the container with id a542b212cf1a53f9b3d07614d5ee19349ca198ceeddee29d9e8276b05190e505 Feb 22 00:12:49 crc kubenswrapper[4857]: I0222 00:12:49.284966 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nx245"] Feb 22 00:12:49 crc kubenswrapper[4857]: W0222 00:12:49.290551 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod361dcb13_b990_403e_814b_cf0d11668bbf.slice/crio-f7e4d3f41f59af1ddd2d75f902412f92ed0f3402fadfe10ee5ba7b9335c63e17 WatchSource:0}: Error finding container f7e4d3f41f59af1ddd2d75f902412f92ed0f3402fadfe10ee5ba7b9335c63e17: Status 404 returned error can't find the container with id f7e4d3f41f59af1ddd2d75f902412f92ed0f3402fadfe10ee5ba7b9335c63e17 Feb 22 00:12:50 crc kubenswrapper[4857]: I0222 00:12:50.097777 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nx245" event={"ID":"361dcb13-b990-403e-814b-cf0d11668bbf","Type":"ContainerStarted","Data":"f7e4d3f41f59af1ddd2d75f902412f92ed0f3402fadfe10ee5ba7b9335c63e17"} Feb 22 00:12:50 crc kubenswrapper[4857]: I0222 00:12:50.098549 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fwkxt" event={"ID":"dc489b84-7090-4294-84f5-77edb82e6ca0","Type":"ContainerStarted","Data":"a542b212cf1a53f9b3d07614d5ee19349ca198ceeddee29d9e8276b05190e505"} Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.106285 4857 generic.go:334] "Generic (PLEG): container finished" podID="b9df334b-3beb-41e4-9ac7-cfaaa10ad923" containerID="e6853a190b1967cfadb93267bb68a5aab36f1c20e978d036591be1cac9359404" exitCode=0 Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.106399 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd7ww" event={"ID":"b9df334b-3beb-41e4-9ac7-cfaaa10ad923","Type":"ContainerDied","Data":"e6853a190b1967cfadb93267bb68a5aab36f1c20e978d036591be1cac9359404"} Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.108442 4857 generic.go:334] "Generic (PLEG): container finished" podID="361dcb13-b990-403e-814b-cf0d11668bbf" containerID="efa2789b014874266ed5bb98d761325c69a0edc23a207852bec83e2878f387f3" exitCode=0 Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.108518 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nx245" event={"ID":"361dcb13-b990-403e-814b-cf0d11668bbf","Type":"ContainerDied","Data":"efa2789b014874266ed5bb98d761325c69a0edc23a207852bec83e2878f387f3"} Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.112839 4857 generic.go:334] "Generic (PLEG): container finished" podID="dc489b84-7090-4294-84f5-77edb82e6ca0" containerID="6285a74b6cd0e19409f2d049cb354ce592b0185449e4cabb1017b51a2c15546a" exitCode=0 Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.112915 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fwkxt" event={"ID":"dc489b84-7090-4294-84f5-77edb82e6ca0","Type":"ContainerDied","Data":"6285a74b6cd0e19409f2d049cb354ce592b0185449e4cabb1017b51a2c15546a"} Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.115851 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xrdpk" event={"ID":"ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2","Type":"ContainerStarted","Data":"1b4a511521de387c6c4ae6c16c37ec1fea445b834363a91ec334b75a95f4953a"} Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.726101 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-7qpsm"] Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.727463 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.741772 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-7qpsm"] Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.829463 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.829547 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9644052-2ee2-4255-a00c-83ca257af59c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.829590 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a9644052-2ee2-4255-a00c-83ca257af59c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.829622 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9644052-2ee2-4255-a00c-83ca257af59c-registry-tls\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.829852 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9644052-2ee2-4255-a00c-83ca257af59c-trusted-ca\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.829920 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86wh8\" (UniqueName: \"kubernetes.io/projected/a9644052-2ee2-4255-a00c-83ca257af59c-kube-api-access-86wh8\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.829986 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a9644052-2ee2-4255-a00c-83ca257af59c-registry-certificates\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.830067 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a9644052-2ee2-4255-a00c-83ca257af59c-bound-sa-token\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.848967 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.931344 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9644052-2ee2-4255-a00c-83ca257af59c-trusted-ca\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.931403 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86wh8\" (UniqueName: \"kubernetes.io/projected/a9644052-2ee2-4255-a00c-83ca257af59c-kube-api-access-86wh8\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.931430 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a9644052-2ee2-4255-a00c-83ca257af59c-registry-certificates\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.931465 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a9644052-2ee2-4255-a00c-83ca257af59c-bound-sa-token\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.931512 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9644052-2ee2-4255-a00c-83ca257af59c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.931560 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a9644052-2ee2-4255-a00c-83ca257af59c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.931596 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9644052-2ee2-4255-a00c-83ca257af59c-registry-tls\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.932127 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a9644052-2ee2-4255-a00c-83ca257af59c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.932669 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a9644052-2ee2-4255-a00c-83ca257af59c-registry-certificates\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.932722 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a9644052-2ee2-4255-a00c-83ca257af59c-trusted-ca\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.938611 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a9644052-2ee2-4255-a00c-83ca257af59c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.938773 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a9644052-2ee2-4255-a00c-83ca257af59c-registry-tls\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.948660 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86wh8\" (UniqueName: \"kubernetes.io/projected/a9644052-2ee2-4255-a00c-83ca257af59c-kube-api-access-86wh8\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:51 crc kubenswrapper[4857]: I0222 00:12:51.949591 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a9644052-2ee2-4255-a00c-83ca257af59c-bound-sa-token\") pod \"image-registry-66df7c8f76-7qpsm\" (UID: \"a9644052-2ee2-4255-a00c-83ca257af59c\") " pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:52 crc kubenswrapper[4857]: I0222 00:12:52.043116 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:52 crc kubenswrapper[4857]: I0222 00:12:52.130016 4857 generic.go:334] "Generic (PLEG): container finished" podID="ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2" containerID="1b4a511521de387c6c4ae6c16c37ec1fea445b834363a91ec334b75a95f4953a" exitCode=0 Feb 22 00:12:52 crc kubenswrapper[4857]: I0222 00:12:52.130250 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xrdpk" event={"ID":"ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2","Type":"ContainerDied","Data":"1b4a511521de387c6c4ae6c16c37ec1fea445b834363a91ec334b75a95f4953a"} Feb 22 00:12:52 crc kubenswrapper[4857]: I0222 00:12:52.458961 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-7qpsm"] Feb 22 00:12:52 crc kubenswrapper[4857]: W0222 00:12:52.471963 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9644052_2ee2_4255_a00c_83ca257af59c.slice/crio-46769b0c4be758a35e33a90b2b59a3c45ee4f0311fb263308437dbc918e35c64 WatchSource:0}: Error finding container 46769b0c4be758a35e33a90b2b59a3c45ee4f0311fb263308437dbc918e35c64: Status 404 returned error can't find the container with id 46769b0c4be758a35e33a90b2b59a3c45ee4f0311fb263308437dbc918e35c64 Feb 22 00:12:53 crc kubenswrapper[4857]: I0222 00:12:53.138571 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" event={"ID":"a9644052-2ee2-4255-a00c-83ca257af59c","Type":"ContainerStarted","Data":"46769b0c4be758a35e33a90b2b59a3c45ee4f0311fb263308437dbc918e35c64"} Feb 22 00:12:54 crc kubenswrapper[4857]: I0222 00:12:54.145804 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd7ww" event={"ID":"b9df334b-3beb-41e4-9ac7-cfaaa10ad923","Type":"ContainerStarted","Data":"b1a4296d2502496dab4c717ff5a70b75d2217381c1b603facd111ff7d1ac2630"} Feb 22 00:12:54 crc kubenswrapper[4857]: I0222 00:12:54.147655 4857 generic.go:334] "Generic (PLEG): container finished" podID="361dcb13-b990-403e-814b-cf0d11668bbf" containerID="883489c58d646f083236379507731f8b72663524d2647da7d38f2888cc7dce45" exitCode=0 Feb 22 00:12:54 crc kubenswrapper[4857]: I0222 00:12:54.147708 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nx245" event={"ID":"361dcb13-b990-403e-814b-cf0d11668bbf","Type":"ContainerDied","Data":"883489c58d646f083236379507731f8b72663524d2647da7d38f2888cc7dce45"} Feb 22 00:12:54 crc kubenswrapper[4857]: I0222 00:12:54.150146 4857 generic.go:334] "Generic (PLEG): container finished" podID="dc489b84-7090-4294-84f5-77edb82e6ca0" containerID="141b407b2285062261fbf827e1244cec1ca43cca5df908ad807b64cb5d1ba266" exitCode=0 Feb 22 00:12:54 crc kubenswrapper[4857]: I0222 00:12:54.150193 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fwkxt" event={"ID":"dc489b84-7090-4294-84f5-77edb82e6ca0","Type":"ContainerDied","Data":"141b407b2285062261fbf827e1244cec1ca43cca5df908ad807b64cb5d1ba266"} Feb 22 00:12:54 crc kubenswrapper[4857]: I0222 00:12:54.152339 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xrdpk" event={"ID":"ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2","Type":"ContainerStarted","Data":"9cf134e2b03bbe960398bb76a004469acacff915d5527b2b20479e400a91f6f6"} Feb 22 00:12:54 crc kubenswrapper[4857]: I0222 00:12:54.154515 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" event={"ID":"a9644052-2ee2-4255-a00c-83ca257af59c","Type":"ContainerStarted","Data":"53bec4b35a0a2b79aa8eaffef77ba5c6948cd7a4ceeab47231a4d5fa695c81a7"} Feb 22 00:12:54 crc kubenswrapper[4857]: I0222 00:12:54.155329 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:12:54 crc kubenswrapper[4857]: I0222 00:12:54.165114 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kd7ww" podStartSLOduration=3.195556733 podStartE2EDuration="8.16509979s" podCreationTimestamp="2026-02-22 00:12:46 +0000 UTC" firstStartedPulling="2026-02-22 00:12:48.082752204 +0000 UTC m=+375.721481497" lastFinishedPulling="2026-02-22 00:12:53.052295291 +0000 UTC m=+380.691024554" observedRunningTime="2026-02-22 00:12:54.163535918 +0000 UTC m=+381.802265191" watchObservedRunningTime="2026-02-22 00:12:54.16509979 +0000 UTC m=+381.803829043" Feb 22 00:12:54 crc kubenswrapper[4857]: I0222 00:12:54.206576 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" podStartSLOduration=3.206561863 podStartE2EDuration="3.206561863s" podCreationTimestamp="2026-02-22 00:12:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:12:54.20282097 +0000 UTC m=+381.841550213" watchObservedRunningTime="2026-02-22 00:12:54.206561863 +0000 UTC m=+381.845291116" Feb 22 00:12:54 crc kubenswrapper[4857]: I0222 00:12:54.239135 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xrdpk" podStartSLOduration=3.135416166 podStartE2EDuration="8.239109093s" podCreationTimestamp="2026-02-22 00:12:46 +0000 UTC" firstStartedPulling="2026-02-22 00:12:48.081621308 +0000 UTC m=+375.720350561" lastFinishedPulling="2026-02-22 00:12:53.185314225 +0000 UTC m=+380.824043488" observedRunningTime="2026-02-22 00:12:54.238025088 +0000 UTC m=+381.876754361" watchObservedRunningTime="2026-02-22 00:12:54.239109093 +0000 UTC m=+381.877838396" Feb 22 00:12:56 crc kubenswrapper[4857]: I0222 00:12:56.166111 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fwkxt" event={"ID":"dc489b84-7090-4294-84f5-77edb82e6ca0","Type":"ContainerStarted","Data":"41346da52f5c00ad193ea3197d0a6b3b85c8fb8bdd50d0eb5f9b6dd2b5d55c2b"} Feb 22 00:12:56 crc kubenswrapper[4857]: I0222 00:12:56.168003 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nx245" event={"ID":"361dcb13-b990-403e-814b-cf0d11668bbf","Type":"ContainerStarted","Data":"d232ef98cc54859a5c185ae89a9a9a0c7ebe24e8efea9a91fbd7803fc34cad96"} Feb 22 00:12:56 crc kubenswrapper[4857]: I0222 00:12:56.185444 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fwkxt" podStartSLOduration=4.234258294 podStartE2EDuration="8.185428248s" podCreationTimestamp="2026-02-22 00:12:48 +0000 UTC" firstStartedPulling="2026-02-22 00:12:51.115233141 +0000 UTC m=+378.753962434" lastFinishedPulling="2026-02-22 00:12:55.066403125 +0000 UTC m=+382.705132388" observedRunningTime="2026-02-22 00:12:56.181971744 +0000 UTC m=+383.820700997" watchObservedRunningTime="2026-02-22 00:12:56.185428248 +0000 UTC m=+383.824157501" Feb 22 00:12:56 crc kubenswrapper[4857]: I0222 00:12:56.205371 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nx245" podStartSLOduration=3.628618202 podStartE2EDuration="8.205349303s" podCreationTimestamp="2026-02-22 00:12:48 +0000 UTC" firstStartedPulling="2026-02-22 00:12:51.110072502 +0000 UTC m=+378.748801755" lastFinishedPulling="2026-02-22 00:12:55.686803603 +0000 UTC m=+383.325532856" observedRunningTime="2026-02-22 00:12:56.198143826 +0000 UTC m=+383.836873089" watchObservedRunningTime="2026-02-22 00:12:56.205349303 +0000 UTC m=+383.844078556" Feb 22 00:12:56 crc kubenswrapper[4857]: I0222 00:12:56.486605 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kd7ww" Feb 22 00:12:56 crc kubenswrapper[4857]: I0222 00:12:56.486677 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kd7ww" Feb 22 00:12:56 crc kubenswrapper[4857]: I0222 00:12:56.524615 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kd7ww" Feb 22 00:12:56 crc kubenswrapper[4857]: I0222 00:12:56.710165 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xrdpk" Feb 22 00:12:56 crc kubenswrapper[4857]: I0222 00:12:56.710219 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xrdpk" Feb 22 00:12:57 crc kubenswrapper[4857]: I0222 00:12:57.757190 4857 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xrdpk" podUID="ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2" containerName="registry-server" probeResult="failure" output=< Feb 22 00:12:57 crc kubenswrapper[4857]: timeout: failed to connect service ":50051" within 1s Feb 22 00:12:57 crc kubenswrapper[4857]: > Feb 22 00:12:58 crc kubenswrapper[4857]: I0222 00:12:58.995231 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fwkxt" Feb 22 00:12:58 crc kubenswrapper[4857]: I0222 00:12:58.995538 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fwkxt" Feb 22 00:12:59 crc kubenswrapper[4857]: I0222 00:12:59.050873 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fwkxt" Feb 22 00:12:59 crc kubenswrapper[4857]: I0222 00:12:59.089864 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nx245" Feb 22 00:12:59 crc kubenswrapper[4857]: I0222 00:12:59.089943 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nx245" Feb 22 00:12:59 crc kubenswrapper[4857]: I0222 00:12:59.132137 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nx245" Feb 22 00:13:06 crc kubenswrapper[4857]: I0222 00:13:06.535593 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kd7ww" Feb 22 00:13:06 crc kubenswrapper[4857]: I0222 00:13:06.772877 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xrdpk" Feb 22 00:13:06 crc kubenswrapper[4857]: I0222 00:13:06.813265 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xrdpk" Feb 22 00:13:09 crc kubenswrapper[4857]: I0222 00:13:09.071190 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fwkxt" Feb 22 00:13:09 crc kubenswrapper[4857]: I0222 00:13:09.143306 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nx245" Feb 22 00:13:10 crc kubenswrapper[4857]: I0222 00:13:10.428794 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:13:10 crc kubenswrapper[4857]: I0222 00:13:10.429078 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:13:10 crc kubenswrapper[4857]: I0222 00:13:10.429118 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:13:10 crc kubenswrapper[4857]: I0222 00:13:10.429585 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93a06a7512e609878c228069a0a49f796632485175ab77273f65221f6e526dd8"} pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 22 00:13:10 crc kubenswrapper[4857]: I0222 00:13:10.429633 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" containerID="cri-o://93a06a7512e609878c228069a0a49f796632485175ab77273f65221f6e526dd8" gracePeriod=600 Feb 22 00:13:11 crc kubenswrapper[4857]: I0222 00:13:11.255932 4857 generic.go:334] "Generic (PLEG): container finished" podID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerID="93a06a7512e609878c228069a0a49f796632485175ab77273f65221f6e526dd8" exitCode=0 Feb 22 00:13:11 crc kubenswrapper[4857]: I0222 00:13:11.256073 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" event={"ID":"2ed2fde8-1449-491b-ae21-2fe11ab008ae","Type":"ContainerDied","Data":"93a06a7512e609878c228069a0a49f796632485175ab77273f65221f6e526dd8"} Feb 22 00:13:11 crc kubenswrapper[4857]: I0222 00:13:11.256488 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" event={"ID":"2ed2fde8-1449-491b-ae21-2fe11ab008ae","Type":"ContainerStarted","Data":"833ca5f17cd1eda283100030bd186ddd6d27df0c724e366f9f63df7686e64ad5"} Feb 22 00:13:11 crc kubenswrapper[4857]: I0222 00:13:11.256539 4857 scope.go:117] "RemoveContainer" containerID="bbaba242522ed04866cc3275766b153ae8712dab0d76a15603bd00305fce72b7" Feb 22 00:13:12 crc kubenswrapper[4857]: I0222 00:13:12.050234 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-7qpsm" Feb 22 00:13:12 crc kubenswrapper[4857]: I0222 00:13:12.102690 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-96hns"] Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.146062 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-96hns" podUID="9e0d9776-fb39-4996-9d7d-48a045cd2da9" containerName="registry" containerID="cri-o://e27f63e562967698d8fca9209ff752f1c5af30fe5cf38c142dde56d26e50bf53" gracePeriod=30 Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.416402 4857 generic.go:334] "Generic (PLEG): container finished" podID="9e0d9776-fb39-4996-9d7d-48a045cd2da9" containerID="e27f63e562967698d8fca9209ff752f1c5af30fe5cf38c142dde56d26e50bf53" exitCode=0 Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.416494 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-96hns" event={"ID":"9e0d9776-fb39-4996-9d7d-48a045cd2da9","Type":"ContainerDied","Data":"e27f63e562967698d8fca9209ff752f1c5af30fe5cf38c142dde56d26e50bf53"} Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.569628 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.647055 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9e0d9776-fb39-4996-9d7d-48a045cd2da9-installation-pull-secrets\") pod \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.647139 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mjr6\" (UniqueName: \"kubernetes.io/projected/9e0d9776-fb39-4996-9d7d-48a045cd2da9-kube-api-access-6mjr6\") pod \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.647277 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.647313 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9e0d9776-fb39-4996-9d7d-48a045cd2da9-registry-tls\") pod \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.647335 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9e0d9776-fb39-4996-9d7d-48a045cd2da9-ca-trust-extracted\") pod \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.647355 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9e0d9776-fb39-4996-9d7d-48a045cd2da9-bound-sa-token\") pod \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.647379 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e0d9776-fb39-4996-9d7d-48a045cd2da9-trusted-ca\") pod \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.647403 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9e0d9776-fb39-4996-9d7d-48a045cd2da9-registry-certificates\") pod \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\" (UID: \"9e0d9776-fb39-4996-9d7d-48a045cd2da9\") " Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.648085 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e0d9776-fb39-4996-9d7d-48a045cd2da9-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "9e0d9776-fb39-4996-9d7d-48a045cd2da9" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.648689 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e0d9776-fb39-4996-9d7d-48a045cd2da9-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9e0d9776-fb39-4996-9d7d-48a045cd2da9" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.653626 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e0d9776-fb39-4996-9d7d-48a045cd2da9-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "9e0d9776-fb39-4996-9d7d-48a045cd2da9" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.654485 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e0d9776-fb39-4996-9d7d-48a045cd2da9-kube-api-access-6mjr6" (OuterVolumeSpecName: "kube-api-access-6mjr6") pod "9e0d9776-fb39-4996-9d7d-48a045cd2da9" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9"). InnerVolumeSpecName "kube-api-access-6mjr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.654660 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e0d9776-fb39-4996-9d7d-48a045cd2da9-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "9e0d9776-fb39-4996-9d7d-48a045cd2da9" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.655906 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e0d9776-fb39-4996-9d7d-48a045cd2da9-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "9e0d9776-fb39-4996-9d7d-48a045cd2da9" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.661149 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "9e0d9776-fb39-4996-9d7d-48a045cd2da9" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.664764 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e0d9776-fb39-4996-9d7d-48a045cd2da9-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "9e0d9776-fb39-4996-9d7d-48a045cd2da9" (UID: "9e0d9776-fb39-4996-9d7d-48a045cd2da9"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.748680 4857 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9e0d9776-fb39-4996-9d7d-48a045cd2da9-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.748732 4857 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e0d9776-fb39-4996-9d7d-48a045cd2da9-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.748744 4857 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9e0d9776-fb39-4996-9d7d-48a045cd2da9-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.748759 4857 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9e0d9776-fb39-4996-9d7d-48a045cd2da9-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.748770 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mjr6\" (UniqueName: \"kubernetes.io/projected/9e0d9776-fb39-4996-9d7d-48a045cd2da9-kube-api-access-6mjr6\") on node \"crc\" DevicePath \"\"" Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.748781 4857 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9e0d9776-fb39-4996-9d7d-48a045cd2da9-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 22 00:13:37 crc kubenswrapper[4857]: I0222 00:13:37.748793 4857 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9e0d9776-fb39-4996-9d7d-48a045cd2da9-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 22 00:13:38 crc kubenswrapper[4857]: I0222 00:13:38.426879 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-96hns" event={"ID":"9e0d9776-fb39-4996-9d7d-48a045cd2da9","Type":"ContainerDied","Data":"502c0c4366db02ccc0c067d724cdcc28c9fcc436a802b1579f1cc78658e56c82"} Feb 22 00:13:38 crc kubenswrapper[4857]: I0222 00:13:38.426934 4857 scope.go:117] "RemoveContainer" containerID="e27f63e562967698d8fca9209ff752f1c5af30fe5cf38c142dde56d26e50bf53" Feb 22 00:13:38 crc kubenswrapper[4857]: I0222 00:13:38.427098 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-96hns" Feb 22 00:13:38 crc kubenswrapper[4857]: I0222 00:13:38.468196 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-96hns"] Feb 22 00:13:38 crc kubenswrapper[4857]: I0222 00:13:38.471693 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-96hns"] Feb 22 00:13:39 crc kubenswrapper[4857]: I0222 00:13:39.083993 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e0d9776-fb39-4996-9d7d-48a045cd2da9" path="/var/lib/kubelet/pods/9e0d9776-fb39-4996-9d7d-48a045cd2da9/volumes" Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.183303 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8"] Feb 22 00:15:00 crc kubenswrapper[4857]: E0222 00:15:00.184189 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e0d9776-fb39-4996-9d7d-48a045cd2da9" containerName="registry" Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.184204 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e0d9776-fb39-4996-9d7d-48a045cd2da9" containerName="registry" Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.184336 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e0d9776-fb39-4996-9d7d-48a045cd2da9" containerName="registry" Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.184848 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8" Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.188427 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.188693 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.189434 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8"] Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.367953 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86v46\" (UniqueName: \"kubernetes.io/projected/a0fc987a-8515-4b5d-97c0-8d0dd8f362eb-kube-api-access-86v46\") pod \"collect-profiles-29528655-m96t8\" (UID: \"a0fc987a-8515-4b5d-97c0-8d0dd8f362eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8" Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.368121 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0fc987a-8515-4b5d-97c0-8d0dd8f362eb-secret-volume\") pod \"collect-profiles-29528655-m96t8\" (UID: \"a0fc987a-8515-4b5d-97c0-8d0dd8f362eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8" Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.368168 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0fc987a-8515-4b5d-97c0-8d0dd8f362eb-config-volume\") pod \"collect-profiles-29528655-m96t8\" (UID: \"a0fc987a-8515-4b5d-97c0-8d0dd8f362eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8" Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.469771 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86v46\" (UniqueName: \"kubernetes.io/projected/a0fc987a-8515-4b5d-97c0-8d0dd8f362eb-kube-api-access-86v46\") pod \"collect-profiles-29528655-m96t8\" (UID: \"a0fc987a-8515-4b5d-97c0-8d0dd8f362eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8" Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.469819 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0fc987a-8515-4b5d-97c0-8d0dd8f362eb-secret-volume\") pod \"collect-profiles-29528655-m96t8\" (UID: \"a0fc987a-8515-4b5d-97c0-8d0dd8f362eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8" Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.469858 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0fc987a-8515-4b5d-97c0-8d0dd8f362eb-config-volume\") pod \"collect-profiles-29528655-m96t8\" (UID: \"a0fc987a-8515-4b5d-97c0-8d0dd8f362eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8" Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.470706 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0fc987a-8515-4b5d-97c0-8d0dd8f362eb-config-volume\") pod \"collect-profiles-29528655-m96t8\" (UID: \"a0fc987a-8515-4b5d-97c0-8d0dd8f362eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8" Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.476438 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0fc987a-8515-4b5d-97c0-8d0dd8f362eb-secret-volume\") pod \"collect-profiles-29528655-m96t8\" (UID: \"a0fc987a-8515-4b5d-97c0-8d0dd8f362eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8" Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.489597 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86v46\" (UniqueName: \"kubernetes.io/projected/a0fc987a-8515-4b5d-97c0-8d0dd8f362eb-kube-api-access-86v46\") pod \"collect-profiles-29528655-m96t8\" (UID: \"a0fc987a-8515-4b5d-97c0-8d0dd8f362eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8" Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.509891 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8" Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.677603 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8"] Feb 22 00:15:00 crc kubenswrapper[4857]: I0222 00:15:00.896702 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8" event={"ID":"a0fc987a-8515-4b5d-97c0-8d0dd8f362eb","Type":"ContainerStarted","Data":"8bacd8950a3d6420340b8b8dcfa1c968c8f74b132b5bb8d9bd74ef886d427e01"} Feb 22 00:15:01 crc kubenswrapper[4857]: I0222 00:15:01.907174 4857 generic.go:334] "Generic (PLEG): container finished" podID="a0fc987a-8515-4b5d-97c0-8d0dd8f362eb" containerID="7e6fd57756f50cf67ef2e31aaad92c15efd76f25ba99560a09401a96a8c6e4ca" exitCode=0 Feb 22 00:15:01 crc kubenswrapper[4857]: I0222 00:15:01.907362 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8" event={"ID":"a0fc987a-8515-4b5d-97c0-8d0dd8f362eb","Type":"ContainerDied","Data":"7e6fd57756f50cf67ef2e31aaad92c15efd76f25ba99560a09401a96a8c6e4ca"} Feb 22 00:15:03 crc kubenswrapper[4857]: I0222 00:15:03.125276 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8" Feb 22 00:15:03 crc kubenswrapper[4857]: I0222 00:15:03.209713 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0fc987a-8515-4b5d-97c0-8d0dd8f362eb-secret-volume\") pod \"a0fc987a-8515-4b5d-97c0-8d0dd8f362eb\" (UID: \"a0fc987a-8515-4b5d-97c0-8d0dd8f362eb\") " Feb 22 00:15:03 crc kubenswrapper[4857]: I0222 00:15:03.210117 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86v46\" (UniqueName: \"kubernetes.io/projected/a0fc987a-8515-4b5d-97c0-8d0dd8f362eb-kube-api-access-86v46\") pod \"a0fc987a-8515-4b5d-97c0-8d0dd8f362eb\" (UID: \"a0fc987a-8515-4b5d-97c0-8d0dd8f362eb\") " Feb 22 00:15:03 crc kubenswrapper[4857]: I0222 00:15:03.210146 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0fc987a-8515-4b5d-97c0-8d0dd8f362eb-config-volume\") pod \"a0fc987a-8515-4b5d-97c0-8d0dd8f362eb\" (UID: \"a0fc987a-8515-4b5d-97c0-8d0dd8f362eb\") " Feb 22 00:15:03 crc kubenswrapper[4857]: I0222 00:15:03.210840 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0fc987a-8515-4b5d-97c0-8d0dd8f362eb-config-volume" (OuterVolumeSpecName: "config-volume") pod "a0fc987a-8515-4b5d-97c0-8d0dd8f362eb" (UID: "a0fc987a-8515-4b5d-97c0-8d0dd8f362eb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:15:03 crc kubenswrapper[4857]: I0222 00:15:03.214914 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0fc987a-8515-4b5d-97c0-8d0dd8f362eb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a0fc987a-8515-4b5d-97c0-8d0dd8f362eb" (UID: "a0fc987a-8515-4b5d-97c0-8d0dd8f362eb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:15:03 crc kubenswrapper[4857]: I0222 00:15:03.215192 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0fc987a-8515-4b5d-97c0-8d0dd8f362eb-kube-api-access-86v46" (OuterVolumeSpecName: "kube-api-access-86v46") pod "a0fc987a-8515-4b5d-97c0-8d0dd8f362eb" (UID: "a0fc987a-8515-4b5d-97c0-8d0dd8f362eb"). InnerVolumeSpecName "kube-api-access-86v46". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:15:03 crc kubenswrapper[4857]: I0222 00:15:03.310936 4857 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a0fc987a-8515-4b5d-97c0-8d0dd8f362eb-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:03 crc kubenswrapper[4857]: I0222 00:15:03.310999 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86v46\" (UniqueName: \"kubernetes.io/projected/a0fc987a-8515-4b5d-97c0-8d0dd8f362eb-kube-api-access-86v46\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:03 crc kubenswrapper[4857]: I0222 00:15:03.311027 4857 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a0fc987a-8515-4b5d-97c0-8d0dd8f362eb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:03 crc kubenswrapper[4857]: I0222 00:15:03.921412 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8" event={"ID":"a0fc987a-8515-4b5d-97c0-8d0dd8f362eb","Type":"ContainerDied","Data":"8bacd8950a3d6420340b8b8dcfa1c968c8f74b132b5bb8d9bd74ef886d427e01"} Feb 22 00:15:03 crc kubenswrapper[4857]: I0222 00:15:03.921461 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8bacd8950a3d6420340b8b8dcfa1c968c8f74b132b5bb8d9bd74ef886d427e01" Feb 22 00:15:03 crc kubenswrapper[4857]: I0222 00:15:03.921516 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29528655-m96t8" Feb 22 00:15:10 crc kubenswrapper[4857]: I0222 00:15:10.428664 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:15:10 crc kubenswrapper[4857]: I0222 00:15:10.429438 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:15:40 crc kubenswrapper[4857]: I0222 00:15:40.428241 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:15:40 crc kubenswrapper[4857]: I0222 00:15:40.428720 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.376977 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6rc8w"] Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.377775 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovn-controller" containerID="cri-o://2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b" gracePeriod=30 Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.377840 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="nbdb" containerID="cri-o://77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2" gracePeriod=30 Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.377910 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovn-acl-logging" containerID="cri-o://38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770" gracePeriod=30 Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.377921 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93" gracePeriod=30 Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.377990 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="northd" containerID="cri-o://18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd" gracePeriod=30 Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.377974 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="kube-rbac-proxy-node" containerID="cri-o://f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9" gracePeriod=30 Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.377918 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="sbdb" containerID="cri-o://a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364" gracePeriod=30 Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.408666 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovnkube-controller" containerID="cri-o://a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366" gracePeriod=30 Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.495183 4857 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.495316 4857 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.497261 4857 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.497383 4857 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.499801 4857 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.499864 4857 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="nbdb" Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.504236 4857 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.504311 4857 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="sbdb" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.693161 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovnkube-controller/3.log" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.694888 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovn-acl-logging/0.log" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.695262 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovn-controller/0.log" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.695581 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741005 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ctw5k"] Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.741206 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovnkube-controller" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741218 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovnkube-controller" Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.741226 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="kubecfg-setup" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741231 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="kubecfg-setup" Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.741237 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovnkube-controller" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741243 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovnkube-controller" Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.741251 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovnkube-controller" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741257 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovnkube-controller" Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.741266 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="kube-rbac-proxy-node" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741271 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="kube-rbac-proxy-node" Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.741285 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovn-controller" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741292 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovn-controller" Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.741301 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="northd" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741308 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="northd" Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.741319 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="sbdb" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741326 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="sbdb" Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.741338 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0fc987a-8515-4b5d-97c0-8d0dd8f362eb" containerName="collect-profiles" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741345 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0fc987a-8515-4b5d-97c0-8d0dd8f362eb" containerName="collect-profiles" Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.741356 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="nbdb" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741363 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="nbdb" Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.741374 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovn-acl-logging" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741379 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovn-acl-logging" Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.741387 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="kube-rbac-proxy-ovn-metrics" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741393 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="kube-rbac-proxy-ovn-metrics" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741474 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="northd" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741483 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="kube-rbac-proxy-node" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741492 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="nbdb" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741500 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovnkube-controller" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741508 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovnkube-controller" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741514 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0fc987a-8515-4b5d-97c0-8d0dd8f362eb" containerName="collect-profiles" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741521 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovnkube-controller" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741528 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovnkube-controller" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741534 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovn-acl-logging" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741545 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovn-controller" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741553 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="kube-rbac-proxy-ovn-metrics" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741559 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="sbdb" Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.741662 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovnkube-controller" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741671 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovnkube-controller" Feb 22 00:15:47 crc kubenswrapper[4857]: E0222 00:15:47.741682 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovnkube-controller" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741691 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovnkube-controller" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.741787 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerName="ovnkube-controller" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.743402 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783415 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9a0db241-7fc5-4d71-a8cc-534f4d303883-ovnkube-config\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783458 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-cni-netd\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783489 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-run-ovn\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783517 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-var-lib-openvswitch\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783539 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-slash\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783566 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9a0db241-7fc5-4d71-a8cc-534f4d303883-ovnkube-script-lib\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783581 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9a0db241-7fc5-4d71-a8cc-534f4d303883-ovn-node-metrics-cert\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783580 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783595 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783601 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-node-log\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783652 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-node-log" (OuterVolumeSpecName: "node-log") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783670 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-run-openvswitch\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783687 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783697 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-run-netns\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783708 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783725 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fw59\" (UniqueName: \"kubernetes.io/projected/9a0db241-7fc5-4d71-a8cc-534f4d303883-kube-api-access-7fw59\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783754 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9a0db241-7fc5-4d71-a8cc-534f4d303883-env-overrides\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783790 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-log-socket\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783824 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-kubelet\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783843 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-cni-bin\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783864 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-systemd-units\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783881 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-var-lib-cni-networks-ovn-kubernetes\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783900 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-run-ovn-kubernetes\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783917 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-etc-openvswitch\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783933 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-run-systemd\") pod \"9a0db241-7fc5-4d71-a8cc-534f4d303883\" (UID: \"9a0db241-7fc5-4d71-a8cc-534f4d303883\") " Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784109 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-cni-netd\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784130 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-slash\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783791 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-slash" (OuterVolumeSpecName: "host-slash") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784149 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-run-netns\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783856 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.783889 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784187 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784198 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784211 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784221 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784166 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-run-ovn\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784276 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-systemd-units\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784306 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ce5c548c-ad23-465a-b06f-a909b46dccd5-ovnkube-config\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784333 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-run-systemd\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784015 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a0db241-7fc5-4d71-a8cc-534f4d303883-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784109 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-log-socket" (OuterVolumeSpecName: "log-socket") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784134 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784505 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ce5c548c-ad23-465a-b06f-a909b46dccd5-ovn-node-metrics-cert\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784670 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gds2j\" (UniqueName: \"kubernetes.io/projected/ce5c548c-ad23-465a-b06f-a909b46dccd5-kube-api-access-gds2j\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784720 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ce5c548c-ad23-465a-b06f-a909b46dccd5-env-overrides\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784742 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-node-log\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784756 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-run-openvswitch\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784761 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a0db241-7fc5-4d71-a8cc-534f4d303883-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784778 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-run-ovn-kubernetes\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784804 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-var-lib-openvswitch\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784851 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-kubelet\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784855 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a0db241-7fc5-4d71-a8cc-534f4d303883-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784877 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.784974 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-log-socket\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785002 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-etc-openvswitch\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785070 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ce5c548c-ad23-465a-b06f-a909b46dccd5-ovnkube-script-lib\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785095 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-cni-bin\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785138 4857 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785150 4857 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-slash\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785158 4857 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9a0db241-7fc5-4d71-a8cc-534f4d303883-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785167 4857 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-node-log\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785174 4857 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785182 4857 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785190 4857 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9a0db241-7fc5-4d71-a8cc-534f4d303883-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785198 4857 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-log-socket\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785206 4857 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785214 4857 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785223 4857 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785231 4857 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785241 4857 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785250 4857 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785259 4857 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9a0db241-7fc5-4d71-a8cc-534f4d303883-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785269 4857 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.785278 4857 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.789153 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a0db241-7fc5-4d71-a8cc-534f4d303883-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.789761 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a0db241-7fc5-4d71-a8cc-534f4d303883-kube-api-access-7fw59" (OuterVolumeSpecName: "kube-api-access-7fw59") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "kube-api-access-7fw59". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.796421 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "9a0db241-7fc5-4d71-a8cc-534f4d303883" (UID: "9a0db241-7fc5-4d71-a8cc-534f4d303883"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886125 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-var-lib-openvswitch\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886176 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-kubelet\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886203 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886244 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-log-socket\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886271 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-etc-openvswitch\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886302 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ce5c548c-ad23-465a-b06f-a909b46dccd5-ovnkube-script-lib\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886296 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-var-lib-openvswitch\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886340 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-log-socket\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886303 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-kubelet\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886372 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-cni-bin\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886327 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-cni-bin\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886400 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-etc-openvswitch\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886410 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886418 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-cni-netd\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886442 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-slash\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886456 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-cni-netd\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886469 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-run-netns\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886629 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-run-ovn\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886643 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-run-netns\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886647 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-slash\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886676 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-systemd-units\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886722 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-run-ovn\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886734 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ce5c548c-ad23-465a-b06f-a909b46dccd5-ovnkube-config\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886748 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-systemd-units\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886765 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-run-systemd\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886813 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ce5c548c-ad23-465a-b06f-a909b46dccd5-ovn-node-metrics-cert\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886854 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gds2j\" (UniqueName: \"kubernetes.io/projected/ce5c548c-ad23-465a-b06f-a909b46dccd5-kube-api-access-gds2j\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886892 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-run-systemd\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886893 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-node-log\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886923 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-node-log\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886955 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ce5c548c-ad23-465a-b06f-a909b46dccd5-env-overrides\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.886978 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-run-openvswitch\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.887014 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-run-ovn-kubernetes\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.887089 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-run-openvswitch\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.887181 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ce5c548c-ad23-465a-b06f-a909b46dccd5-ovnkube-script-lib\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.887800 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ce5c548c-ad23-465a-b06f-a909b46dccd5-ovnkube-config\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.888002 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ce5c548c-ad23-465a-b06f-a909b46dccd5-env-overrides\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.890258 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ce5c548c-ad23-465a-b06f-a909b46dccd5-host-run-ovn-kubernetes\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.890359 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fw59\" (UniqueName: \"kubernetes.io/projected/9a0db241-7fc5-4d71-a8cc-534f4d303883-kube-api-access-7fw59\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.890381 4857 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9a0db241-7fc5-4d71-a8cc-534f4d303883-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.890403 4857 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9a0db241-7fc5-4d71-a8cc-534f4d303883-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.898333 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ce5c548c-ad23-465a-b06f-a909b46dccd5-ovn-node-metrics-cert\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:47 crc kubenswrapper[4857]: I0222 00:15:47.913024 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gds2j\" (UniqueName: \"kubernetes.io/projected/ce5c548c-ad23-465a-b06f-a909b46dccd5-kube-api-access-gds2j\") pod \"ovnkube-node-ctw5k\" (UID: \"ce5c548c-ad23-465a-b06f-a909b46dccd5\") " pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.061157 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.149900 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" event={"ID":"ce5c548c-ad23-465a-b06f-a909b46dccd5","Type":"ContainerStarted","Data":"0a12d0e5af461d2bc9bf5dd8ab4525bf858382f14d0c8cae7043e26481d9ad82"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.151543 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4lmlt_9e857d8a-7289-4352-9de9-1b5d0bd21e8f/kube-multus/2.log" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.152061 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4lmlt_9e857d8a-7289-4352-9de9-1b5d0bd21e8f/kube-multus/1.log" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.152103 4857 generic.go:334] "Generic (PLEG): container finished" podID="9e857d8a-7289-4352-9de9-1b5d0bd21e8f" containerID="57944544b90601bf139fb88d074d1c7fc35b6050e563aa238c6f1f02184473a4" exitCode=2 Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.152180 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4lmlt" event={"ID":"9e857d8a-7289-4352-9de9-1b5d0bd21e8f","Type":"ContainerDied","Data":"57944544b90601bf139fb88d074d1c7fc35b6050e563aa238c6f1f02184473a4"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.152516 4857 scope.go:117] "RemoveContainer" containerID="28685fd8939b71352e0be13ea9b6f38671b481eb13c5d08c79e788e8fc1c1f1c" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.152915 4857 scope.go:117] "RemoveContainer" containerID="57944544b90601bf139fb88d074d1c7fc35b6050e563aa238c6f1f02184473a4" Feb 22 00:15:48 crc kubenswrapper[4857]: E0222 00:15:48.153124 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-4lmlt_openshift-multus(9e857d8a-7289-4352-9de9-1b5d0bd21e8f)\"" pod="openshift-multus/multus-4lmlt" podUID="9e857d8a-7289-4352-9de9-1b5d0bd21e8f" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.154158 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovnkube-controller/3.log" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.158380 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovn-acl-logging/0.log" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.158878 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6rc8w_9a0db241-7fc5-4d71-a8cc-534f4d303883/ovn-controller/0.log" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159270 4857 generic.go:334] "Generic (PLEG): container finished" podID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerID="a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366" exitCode=0 Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159294 4857 generic.go:334] "Generic (PLEG): container finished" podID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerID="a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364" exitCode=0 Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159302 4857 generic.go:334] "Generic (PLEG): container finished" podID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerID="77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2" exitCode=0 Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159308 4857 generic.go:334] "Generic (PLEG): container finished" podID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerID="18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd" exitCode=0 Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159317 4857 generic.go:334] "Generic (PLEG): container finished" podID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerID="fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93" exitCode=0 Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159324 4857 generic.go:334] "Generic (PLEG): container finished" podID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerID="f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9" exitCode=0 Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159332 4857 generic.go:334] "Generic (PLEG): container finished" podID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerID="38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770" exitCode=143 Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159336 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159350 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerDied","Data":"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159394 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerDied","Data":"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159413 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerDied","Data":"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159339 4857 generic.go:334] "Generic (PLEG): container finished" podID="9a0db241-7fc5-4d71-a8cc-534f4d303883" containerID="2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b" exitCode=143 Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159424 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerDied","Data":"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159460 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerDied","Data":"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159477 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerDied","Data":"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159488 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159497 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159503 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159508 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159514 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159519 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159523 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159528 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159534 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159538 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159545 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerDied","Data":"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159552 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159558 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159563 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159567 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159572 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159580 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159591 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159601 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159610 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159617 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159627 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerDied","Data":"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159639 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159647 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159653 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159660 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159666 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159673 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159679 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159685 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159691 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159697 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159708 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6rc8w" event={"ID":"9a0db241-7fc5-4d71-a8cc-534f4d303883","Type":"ContainerDied","Data":"01681e4530f602e5e717500243924331ffb7efdd3b4aca046931512101cb8c6f"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159720 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159727 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159733 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159739 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159745 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159752 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159759 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159765 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159772 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.159778 4857 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b"} Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.180251 4857 scope.go:117] "RemoveContainer" containerID="a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.199618 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6rc8w"] Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.202711 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6rc8w"] Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.259979 4857 scope.go:117] "RemoveContainer" containerID="f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.277270 4857 scope.go:117] "RemoveContainer" containerID="a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.299075 4857 scope.go:117] "RemoveContainer" containerID="77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.325023 4857 scope.go:117] "RemoveContainer" containerID="18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.339342 4857 scope.go:117] "RemoveContainer" containerID="fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.351331 4857 scope.go:117] "RemoveContainer" containerID="f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.364274 4857 scope.go:117] "RemoveContainer" containerID="38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.376937 4857 scope.go:117] "RemoveContainer" containerID="2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.391117 4857 scope.go:117] "RemoveContainer" containerID="f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.404867 4857 scope.go:117] "RemoveContainer" containerID="a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366" Feb 22 00:15:48 crc kubenswrapper[4857]: E0222 00:15:48.405426 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366\": container with ID starting with a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366 not found: ID does not exist" containerID="a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.405468 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366"} err="failed to get container status \"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366\": rpc error: code = NotFound desc = could not find container \"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366\": container with ID starting with a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.405498 4857 scope.go:117] "RemoveContainer" containerID="f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715" Feb 22 00:15:48 crc kubenswrapper[4857]: E0222 00:15:48.405866 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\": container with ID starting with f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715 not found: ID does not exist" containerID="f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.405886 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715"} err="failed to get container status \"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\": rpc error: code = NotFound desc = could not find container \"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\": container with ID starting with f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.405899 4857 scope.go:117] "RemoveContainer" containerID="a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364" Feb 22 00:15:48 crc kubenswrapper[4857]: E0222 00:15:48.406291 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\": container with ID starting with a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364 not found: ID does not exist" containerID="a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.406330 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364"} err="failed to get container status \"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\": rpc error: code = NotFound desc = could not find container \"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\": container with ID starting with a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.406363 4857 scope.go:117] "RemoveContainer" containerID="77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2" Feb 22 00:15:48 crc kubenswrapper[4857]: E0222 00:15:48.406673 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\": container with ID starting with 77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2 not found: ID does not exist" containerID="77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.406700 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2"} err="failed to get container status \"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\": rpc error: code = NotFound desc = could not find container \"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\": container with ID starting with 77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.406714 4857 scope.go:117] "RemoveContainer" containerID="18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd" Feb 22 00:15:48 crc kubenswrapper[4857]: E0222 00:15:48.406972 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\": container with ID starting with 18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd not found: ID does not exist" containerID="18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.407019 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd"} err="failed to get container status \"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\": rpc error: code = NotFound desc = could not find container \"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\": container with ID starting with 18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.407067 4857 scope.go:117] "RemoveContainer" containerID="fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93" Feb 22 00:15:48 crc kubenswrapper[4857]: E0222 00:15:48.407344 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\": container with ID starting with fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93 not found: ID does not exist" containerID="fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.407367 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93"} err="failed to get container status \"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\": rpc error: code = NotFound desc = could not find container \"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\": container with ID starting with fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.407381 4857 scope.go:117] "RemoveContainer" containerID="f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9" Feb 22 00:15:48 crc kubenswrapper[4857]: E0222 00:15:48.407583 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\": container with ID starting with f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9 not found: ID does not exist" containerID="f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.407607 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9"} err="failed to get container status \"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\": rpc error: code = NotFound desc = could not find container \"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\": container with ID starting with f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.407624 4857 scope.go:117] "RemoveContainer" containerID="38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770" Feb 22 00:15:48 crc kubenswrapper[4857]: E0222 00:15:48.407858 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\": container with ID starting with 38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770 not found: ID does not exist" containerID="38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.407883 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770"} err="failed to get container status \"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\": rpc error: code = NotFound desc = could not find container \"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\": container with ID starting with 38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.407898 4857 scope.go:117] "RemoveContainer" containerID="2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b" Feb 22 00:15:48 crc kubenswrapper[4857]: E0222 00:15:48.408195 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\": container with ID starting with 2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b not found: ID does not exist" containerID="2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.408216 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b"} err="failed to get container status \"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\": rpc error: code = NotFound desc = could not find container \"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\": container with ID starting with 2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.408227 4857 scope.go:117] "RemoveContainer" containerID="f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b" Feb 22 00:15:48 crc kubenswrapper[4857]: E0222 00:15:48.408496 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\": container with ID starting with f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b not found: ID does not exist" containerID="f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.408521 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b"} err="failed to get container status \"f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\": rpc error: code = NotFound desc = could not find container \"f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\": container with ID starting with f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.408535 4857 scope.go:117] "RemoveContainer" containerID="a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.408785 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366"} err="failed to get container status \"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366\": rpc error: code = NotFound desc = could not find container \"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366\": container with ID starting with a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.408804 4857 scope.go:117] "RemoveContainer" containerID="f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.409115 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715"} err="failed to get container status \"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\": rpc error: code = NotFound desc = could not find container \"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\": container with ID starting with f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.409133 4857 scope.go:117] "RemoveContainer" containerID="a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.409389 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364"} err="failed to get container status \"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\": rpc error: code = NotFound desc = could not find container \"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\": container with ID starting with a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.409414 4857 scope.go:117] "RemoveContainer" containerID="77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.409750 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2"} err="failed to get container status \"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\": rpc error: code = NotFound desc = could not find container \"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\": container with ID starting with 77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.409770 4857 scope.go:117] "RemoveContainer" containerID="18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.410110 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd"} err="failed to get container status \"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\": rpc error: code = NotFound desc = could not find container \"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\": container with ID starting with 18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.410138 4857 scope.go:117] "RemoveContainer" containerID="fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.410407 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93"} err="failed to get container status \"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\": rpc error: code = NotFound desc = could not find container \"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\": container with ID starting with fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.410426 4857 scope.go:117] "RemoveContainer" containerID="f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.410646 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9"} err="failed to get container status \"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\": rpc error: code = NotFound desc = could not find container \"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\": container with ID starting with f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.410673 4857 scope.go:117] "RemoveContainer" containerID="38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.410909 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770"} err="failed to get container status \"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\": rpc error: code = NotFound desc = could not find container \"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\": container with ID starting with 38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.410947 4857 scope.go:117] "RemoveContainer" containerID="2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.411193 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b"} err="failed to get container status \"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\": rpc error: code = NotFound desc = could not find container \"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\": container with ID starting with 2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.411224 4857 scope.go:117] "RemoveContainer" containerID="f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.411538 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b"} err="failed to get container status \"f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\": rpc error: code = NotFound desc = could not find container \"f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\": container with ID starting with f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.411568 4857 scope.go:117] "RemoveContainer" containerID="a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.411819 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366"} err="failed to get container status \"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366\": rpc error: code = NotFound desc = could not find container \"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366\": container with ID starting with a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.411841 4857 scope.go:117] "RemoveContainer" containerID="f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.412160 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715"} err="failed to get container status \"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\": rpc error: code = NotFound desc = could not find container \"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\": container with ID starting with f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.412179 4857 scope.go:117] "RemoveContainer" containerID="a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.412437 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364"} err="failed to get container status \"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\": rpc error: code = NotFound desc = could not find container \"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\": container with ID starting with a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.412456 4857 scope.go:117] "RemoveContainer" containerID="77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.412702 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2"} err="failed to get container status \"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\": rpc error: code = NotFound desc = could not find container \"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\": container with ID starting with 77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.412721 4857 scope.go:117] "RemoveContainer" containerID="18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.412934 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd"} err="failed to get container status \"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\": rpc error: code = NotFound desc = could not find container \"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\": container with ID starting with 18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.412977 4857 scope.go:117] "RemoveContainer" containerID="fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.413335 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93"} err="failed to get container status \"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\": rpc error: code = NotFound desc = could not find container \"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\": container with ID starting with fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.413384 4857 scope.go:117] "RemoveContainer" containerID="f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.413617 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9"} err="failed to get container status \"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\": rpc error: code = NotFound desc = could not find container \"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\": container with ID starting with f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.413639 4857 scope.go:117] "RemoveContainer" containerID="38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.413930 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770"} err="failed to get container status \"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\": rpc error: code = NotFound desc = could not find container \"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\": container with ID starting with 38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.413951 4857 scope.go:117] "RemoveContainer" containerID="2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.414166 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b"} err="failed to get container status \"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\": rpc error: code = NotFound desc = could not find container \"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\": container with ID starting with 2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.414193 4857 scope.go:117] "RemoveContainer" containerID="f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.414461 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b"} err="failed to get container status \"f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\": rpc error: code = NotFound desc = could not find container \"f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\": container with ID starting with f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.414481 4857 scope.go:117] "RemoveContainer" containerID="a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.414730 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366"} err="failed to get container status \"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366\": rpc error: code = NotFound desc = could not find container \"a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366\": container with ID starting with a3660cfc9e8c08ae24a4e8d8b13a72285983d64e5dcd9937ae9ba7246ea71366 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.414752 4857 scope.go:117] "RemoveContainer" containerID="f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.414989 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715"} err="failed to get container status \"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\": rpc error: code = NotFound desc = could not find container \"f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715\": container with ID starting with f5d4e5bbfacea5f91bd1732b53bd508eaa7b9619b0155517486783b858b5b715 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.415007 4857 scope.go:117] "RemoveContainer" containerID="a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.415273 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364"} err="failed to get container status \"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\": rpc error: code = NotFound desc = could not find container \"a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364\": container with ID starting with a5ab6b7fe5ac2142261c9fbf8e51ad0127bd2d48adfc3b40e39fad850494d364 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.415290 4857 scope.go:117] "RemoveContainer" containerID="77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.415535 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2"} err="failed to get container status \"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\": rpc error: code = NotFound desc = could not find container \"77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2\": container with ID starting with 77f3a9755b33a96f1d097b87e9e6e35f15a6764b60354788af1cc375caca4bc2 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.415556 4857 scope.go:117] "RemoveContainer" containerID="18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.415786 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd"} err="failed to get container status \"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\": rpc error: code = NotFound desc = could not find container \"18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd\": container with ID starting with 18900fd08d15f64aa2745fe7e3ff0baa5761ff754f419b20bc5f7df692b2e9bd not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.415804 4857 scope.go:117] "RemoveContainer" containerID="fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.416075 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93"} err="failed to get container status \"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\": rpc error: code = NotFound desc = could not find container \"fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93\": container with ID starting with fda19a286e14ade376c29959234fb620be180d2612c5f64f616c0e115041ae93 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.416093 4857 scope.go:117] "RemoveContainer" containerID="f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.416370 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9"} err="failed to get container status \"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\": rpc error: code = NotFound desc = could not find container \"f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9\": container with ID starting with f1c5d328bf116bc2b41d071a026a8cba3ff2087ab61cab70b12a00a104b9f3e9 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.416395 4857 scope.go:117] "RemoveContainer" containerID="38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.416710 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770"} err="failed to get container status \"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\": rpc error: code = NotFound desc = could not find container \"38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770\": container with ID starting with 38b662bcdeda27609c23f3788d32dd5d30a1d75fe3b57453533e52d980bdf770 not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.416730 4857 scope.go:117] "RemoveContainer" containerID="2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.416944 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b"} err="failed to get container status \"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\": rpc error: code = NotFound desc = could not find container \"2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b\": container with ID starting with 2f10f8468b549b61fc870ece712d06254b66797989ce34d6613e1effb1c6003b not found: ID does not exist" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.416966 4857 scope.go:117] "RemoveContainer" containerID="f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b" Feb 22 00:15:48 crc kubenswrapper[4857]: I0222 00:15:48.417246 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b"} err="failed to get container status \"f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\": rpc error: code = NotFound desc = could not find container \"f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b\": container with ID starting with f9a1802ff9cc3bdd53c40f597c04155e634e599afdc4029696cb18008d38170b not found: ID does not exist" Feb 22 00:15:49 crc kubenswrapper[4857]: I0222 00:15:49.084786 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a0db241-7fc5-4d71-a8cc-534f4d303883" path="/var/lib/kubelet/pods/9a0db241-7fc5-4d71-a8cc-534f4d303883/volumes" Feb 22 00:15:49 crc kubenswrapper[4857]: I0222 00:15:49.168907 4857 generic.go:334] "Generic (PLEG): container finished" podID="ce5c548c-ad23-465a-b06f-a909b46dccd5" containerID="68d228eac932173ee23d64097840272497ca1af6be9d49b41b5411f055b99bee" exitCode=0 Feb 22 00:15:49 crc kubenswrapper[4857]: I0222 00:15:49.169104 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" event={"ID":"ce5c548c-ad23-465a-b06f-a909b46dccd5","Type":"ContainerDied","Data":"68d228eac932173ee23d64097840272497ca1af6be9d49b41b5411f055b99bee"} Feb 22 00:15:49 crc kubenswrapper[4857]: I0222 00:15:49.171239 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4lmlt_9e857d8a-7289-4352-9de9-1b5d0bd21e8f/kube-multus/2.log" Feb 22 00:15:50 crc kubenswrapper[4857]: I0222 00:15:50.182235 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" event={"ID":"ce5c548c-ad23-465a-b06f-a909b46dccd5","Type":"ContainerStarted","Data":"6086565ec28f0a3890506a84bcc2451144c37d6aa44737e22238c06f702b2070"} Feb 22 00:15:50 crc kubenswrapper[4857]: I0222 00:15:50.182566 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" event={"ID":"ce5c548c-ad23-465a-b06f-a909b46dccd5","Type":"ContainerStarted","Data":"e8596fcd50744a8be19dd134c1661848e95560fff66ccdaca4515bcb4f9802a1"} Feb 22 00:15:50 crc kubenswrapper[4857]: I0222 00:15:50.182583 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" event={"ID":"ce5c548c-ad23-465a-b06f-a909b46dccd5","Type":"ContainerStarted","Data":"d96b1b04a65439ebb73deced40ed47c8dded5ff0f7073fa704c78cf4062ad728"} Feb 22 00:15:50 crc kubenswrapper[4857]: I0222 00:15:50.182595 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" event={"ID":"ce5c548c-ad23-465a-b06f-a909b46dccd5","Type":"ContainerStarted","Data":"a6f0720b938274cc7236dc0ba1216350dd4faccb495b4e747c4c93b15d325c1c"} Feb 22 00:15:50 crc kubenswrapper[4857]: I0222 00:15:50.182607 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" event={"ID":"ce5c548c-ad23-465a-b06f-a909b46dccd5","Type":"ContainerStarted","Data":"821332da5311d2d2578c47ce9edea48b968dcfa5e7cf126e1e63efcede87b9e0"} Feb 22 00:15:50 crc kubenswrapper[4857]: I0222 00:15:50.182617 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" event={"ID":"ce5c548c-ad23-465a-b06f-a909b46dccd5","Type":"ContainerStarted","Data":"8a07db95b163c69e0f2bc1daffecd6a94d70949f5a5cb511d9e1cd01971c711d"} Feb 22 00:15:53 crc kubenswrapper[4857]: I0222 00:15:53.201352 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" event={"ID":"ce5c548c-ad23-465a-b06f-a909b46dccd5","Type":"ContainerStarted","Data":"20ffafc6e9b40895b0b8e13a163de609b61d2b4785d83b6ffb7e710db8529bd8"} Feb 22 00:15:55 crc kubenswrapper[4857]: I0222 00:15:55.214818 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" event={"ID":"ce5c548c-ad23-465a-b06f-a909b46dccd5","Type":"ContainerStarted","Data":"0bcaef997f8084c19a2cc39915ee50f50e4b7f41f6bc33339120fe96851c8678"} Feb 22 00:15:55 crc kubenswrapper[4857]: I0222 00:15:55.215573 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:55 crc kubenswrapper[4857]: I0222 00:15:55.215603 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:55 crc kubenswrapper[4857]: I0222 00:15:55.215625 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:55 crc kubenswrapper[4857]: I0222 00:15:55.243836 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:55 crc kubenswrapper[4857]: I0222 00:15:55.245214 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:15:55 crc kubenswrapper[4857]: I0222 00:15:55.255816 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" podStartSLOduration=8.255799484 podStartE2EDuration="8.255799484s" podCreationTimestamp="2026-02-22 00:15:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:15:55.251767385 +0000 UTC m=+562.890496638" watchObservedRunningTime="2026-02-22 00:15:55.255799484 +0000 UTC m=+562.894528737" Feb 22 00:16:03 crc kubenswrapper[4857]: I0222 00:16:03.082409 4857 scope.go:117] "RemoveContainer" containerID="57944544b90601bf139fb88d074d1c7fc35b6050e563aa238c6f1f02184473a4" Feb 22 00:16:03 crc kubenswrapper[4857]: E0222 00:16:03.083225 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-4lmlt_openshift-multus(9e857d8a-7289-4352-9de9-1b5d0bd21e8f)\"" pod="openshift-multus/multus-4lmlt" podUID="9e857d8a-7289-4352-9de9-1b5d0bd21e8f" Feb 22 00:16:10 crc kubenswrapper[4857]: I0222 00:16:10.429003 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:16:10 crc kubenswrapper[4857]: I0222 00:16:10.429417 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:16:10 crc kubenswrapper[4857]: I0222 00:16:10.429465 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:16:10 crc kubenswrapper[4857]: I0222 00:16:10.430012 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"833ca5f17cd1eda283100030bd186ddd6d27df0c724e366f9f63df7686e64ad5"} pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 22 00:16:10 crc kubenswrapper[4857]: I0222 00:16:10.430130 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" containerID="cri-o://833ca5f17cd1eda283100030bd186ddd6d27df0c724e366f9f63df7686e64ad5" gracePeriod=600 Feb 22 00:16:11 crc kubenswrapper[4857]: I0222 00:16:11.305763 4857 generic.go:334] "Generic (PLEG): container finished" podID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerID="833ca5f17cd1eda283100030bd186ddd6d27df0c724e366f9f63df7686e64ad5" exitCode=0 Feb 22 00:16:11 crc kubenswrapper[4857]: I0222 00:16:11.306134 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" event={"ID":"2ed2fde8-1449-491b-ae21-2fe11ab008ae","Type":"ContainerDied","Data":"833ca5f17cd1eda283100030bd186ddd6d27df0c724e366f9f63df7686e64ad5"} Feb 22 00:16:11 crc kubenswrapper[4857]: I0222 00:16:11.306161 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" event={"ID":"2ed2fde8-1449-491b-ae21-2fe11ab008ae","Type":"ContainerStarted","Data":"844b27723312ec8a6f8db792646f1c5f5e12d2eaf2bbdc056197585dfd92820e"} Feb 22 00:16:11 crc kubenswrapper[4857]: I0222 00:16:11.306176 4857 scope.go:117] "RemoveContainer" containerID="93a06a7512e609878c228069a0a49f796632485175ab77273f65221f6e526dd8" Feb 22 00:16:17 crc kubenswrapper[4857]: I0222 00:16:17.077391 4857 scope.go:117] "RemoveContainer" containerID="57944544b90601bf139fb88d074d1c7fc35b6050e563aa238c6f1f02184473a4" Feb 22 00:16:17 crc kubenswrapper[4857]: I0222 00:16:17.354521 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4lmlt_9e857d8a-7289-4352-9de9-1b5d0bd21e8f/kube-multus/2.log" Feb 22 00:16:17 crc kubenswrapper[4857]: I0222 00:16:17.354847 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4lmlt" event={"ID":"9e857d8a-7289-4352-9de9-1b5d0bd21e8f","Type":"ContainerStarted","Data":"a7756cefda2e2f68c7d5aaa1d983e777f86f01f9084ffe1df34f620256098532"} Feb 22 00:16:18 crc kubenswrapper[4857]: I0222 00:16:18.090027 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ctw5k" Feb 22 00:16:50 crc kubenswrapper[4857]: I0222 00:16:50.808721 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd7ww"] Feb 22 00:16:50 crc kubenswrapper[4857]: I0222 00:16:50.809802 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kd7ww" podUID="b9df334b-3beb-41e4-9ac7-cfaaa10ad923" containerName="registry-server" containerID="cri-o://b1a4296d2502496dab4c717ff5a70b75d2217381c1b603facd111ff7d1ac2630" gracePeriod=30 Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.141014 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kd7ww" Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.295903 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9df334b-3beb-41e4-9ac7-cfaaa10ad923-catalog-content\") pod \"b9df334b-3beb-41e4-9ac7-cfaaa10ad923\" (UID: \"b9df334b-3beb-41e4-9ac7-cfaaa10ad923\") " Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.295950 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9df334b-3beb-41e4-9ac7-cfaaa10ad923-utilities\") pod \"b9df334b-3beb-41e4-9ac7-cfaaa10ad923\" (UID: \"b9df334b-3beb-41e4-9ac7-cfaaa10ad923\") " Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.296004 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swfzp\" (UniqueName: \"kubernetes.io/projected/b9df334b-3beb-41e4-9ac7-cfaaa10ad923-kube-api-access-swfzp\") pod \"b9df334b-3beb-41e4-9ac7-cfaaa10ad923\" (UID: \"b9df334b-3beb-41e4-9ac7-cfaaa10ad923\") " Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.297184 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9df334b-3beb-41e4-9ac7-cfaaa10ad923-utilities" (OuterVolumeSpecName: "utilities") pod "b9df334b-3beb-41e4-9ac7-cfaaa10ad923" (UID: "b9df334b-3beb-41e4-9ac7-cfaaa10ad923"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.301521 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9df334b-3beb-41e4-9ac7-cfaaa10ad923-kube-api-access-swfzp" (OuterVolumeSpecName: "kube-api-access-swfzp") pod "b9df334b-3beb-41e4-9ac7-cfaaa10ad923" (UID: "b9df334b-3beb-41e4-9ac7-cfaaa10ad923"). InnerVolumeSpecName "kube-api-access-swfzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.325197 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9df334b-3beb-41e4-9ac7-cfaaa10ad923-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9df334b-3beb-41e4-9ac7-cfaaa10ad923" (UID: "b9df334b-3beb-41e4-9ac7-cfaaa10ad923"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.397128 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9df334b-3beb-41e4-9ac7-cfaaa10ad923-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.397188 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9df334b-3beb-41e4-9ac7-cfaaa10ad923-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.397202 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swfzp\" (UniqueName: \"kubernetes.io/projected/b9df334b-3beb-41e4-9ac7-cfaaa10ad923-kube-api-access-swfzp\") on node \"crc\" DevicePath \"\"" Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.545481 4857 generic.go:334] "Generic (PLEG): container finished" podID="b9df334b-3beb-41e4-9ac7-cfaaa10ad923" containerID="b1a4296d2502496dab4c717ff5a70b75d2217381c1b603facd111ff7d1ac2630" exitCode=0 Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.545567 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kd7ww" Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.545806 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd7ww" event={"ID":"b9df334b-3beb-41e4-9ac7-cfaaa10ad923","Type":"ContainerDied","Data":"b1a4296d2502496dab4c717ff5a70b75d2217381c1b603facd111ff7d1ac2630"} Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.545900 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kd7ww" event={"ID":"b9df334b-3beb-41e4-9ac7-cfaaa10ad923","Type":"ContainerDied","Data":"97c1c99e002c7300d437014c4873e99bca6ec77fbec1972093fa5cf70c4ae7d0"} Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.545967 4857 scope.go:117] "RemoveContainer" containerID="b1a4296d2502496dab4c717ff5a70b75d2217381c1b603facd111ff7d1ac2630" Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.560264 4857 scope.go:117] "RemoveContainer" containerID="e6853a190b1967cfadb93267bb68a5aab36f1c20e978d036591be1cac9359404" Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.572184 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd7ww"] Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.575084 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kd7ww"] Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.576140 4857 scope.go:117] "RemoveContainer" containerID="8e06bf1bbd274d6e83ebe407c102dc09261e04dbf127cebb87acd2ed0c796d52" Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.596671 4857 scope.go:117] "RemoveContainer" containerID="b1a4296d2502496dab4c717ff5a70b75d2217381c1b603facd111ff7d1ac2630" Feb 22 00:16:51 crc kubenswrapper[4857]: E0222 00:16:51.597014 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1a4296d2502496dab4c717ff5a70b75d2217381c1b603facd111ff7d1ac2630\": container with ID starting with b1a4296d2502496dab4c717ff5a70b75d2217381c1b603facd111ff7d1ac2630 not found: ID does not exist" containerID="b1a4296d2502496dab4c717ff5a70b75d2217381c1b603facd111ff7d1ac2630" Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.597070 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1a4296d2502496dab4c717ff5a70b75d2217381c1b603facd111ff7d1ac2630"} err="failed to get container status \"b1a4296d2502496dab4c717ff5a70b75d2217381c1b603facd111ff7d1ac2630\": rpc error: code = NotFound desc = could not find container \"b1a4296d2502496dab4c717ff5a70b75d2217381c1b603facd111ff7d1ac2630\": container with ID starting with b1a4296d2502496dab4c717ff5a70b75d2217381c1b603facd111ff7d1ac2630 not found: ID does not exist" Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.597090 4857 scope.go:117] "RemoveContainer" containerID="e6853a190b1967cfadb93267bb68a5aab36f1c20e978d036591be1cac9359404" Feb 22 00:16:51 crc kubenswrapper[4857]: E0222 00:16:51.597425 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6853a190b1967cfadb93267bb68a5aab36f1c20e978d036591be1cac9359404\": container with ID starting with e6853a190b1967cfadb93267bb68a5aab36f1c20e978d036591be1cac9359404 not found: ID does not exist" containerID="e6853a190b1967cfadb93267bb68a5aab36f1c20e978d036591be1cac9359404" Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.597449 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6853a190b1967cfadb93267bb68a5aab36f1c20e978d036591be1cac9359404"} err="failed to get container status \"e6853a190b1967cfadb93267bb68a5aab36f1c20e978d036591be1cac9359404\": rpc error: code = NotFound desc = could not find container \"e6853a190b1967cfadb93267bb68a5aab36f1c20e978d036591be1cac9359404\": container with ID starting with e6853a190b1967cfadb93267bb68a5aab36f1c20e978d036591be1cac9359404 not found: ID does not exist" Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.597464 4857 scope.go:117] "RemoveContainer" containerID="8e06bf1bbd274d6e83ebe407c102dc09261e04dbf127cebb87acd2ed0c796d52" Feb 22 00:16:51 crc kubenswrapper[4857]: E0222 00:16:51.597673 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e06bf1bbd274d6e83ebe407c102dc09261e04dbf127cebb87acd2ed0c796d52\": container with ID starting with 8e06bf1bbd274d6e83ebe407c102dc09261e04dbf127cebb87acd2ed0c796d52 not found: ID does not exist" containerID="8e06bf1bbd274d6e83ebe407c102dc09261e04dbf127cebb87acd2ed0c796d52" Feb 22 00:16:51 crc kubenswrapper[4857]: I0222 00:16:51.597700 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e06bf1bbd274d6e83ebe407c102dc09261e04dbf127cebb87acd2ed0c796d52"} err="failed to get container status \"8e06bf1bbd274d6e83ebe407c102dc09261e04dbf127cebb87acd2ed0c796d52\": rpc error: code = NotFound desc = could not find container \"8e06bf1bbd274d6e83ebe407c102dc09261e04dbf127cebb87acd2ed0c796d52\": container with ID starting with 8e06bf1bbd274d6e83ebe407c102dc09261e04dbf127cebb87acd2ed0c796d52 not found: ID does not exist" Feb 22 00:16:53 crc kubenswrapper[4857]: I0222 00:16:53.084980 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9df334b-3beb-41e4-9ac7-cfaaa10ad923" path="/var/lib/kubelet/pods/b9df334b-3beb-41e4-9ac7-cfaaa10ad923/volumes" Feb 22 00:16:54 crc kubenswrapper[4857]: I0222 00:16:54.527560 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q"] Feb 22 00:16:54 crc kubenswrapper[4857]: E0222 00:16:54.527807 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9df334b-3beb-41e4-9ac7-cfaaa10ad923" containerName="extract-content" Feb 22 00:16:54 crc kubenswrapper[4857]: I0222 00:16:54.527822 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9df334b-3beb-41e4-9ac7-cfaaa10ad923" containerName="extract-content" Feb 22 00:16:54 crc kubenswrapper[4857]: E0222 00:16:54.527833 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9df334b-3beb-41e4-9ac7-cfaaa10ad923" containerName="registry-server" Feb 22 00:16:54 crc kubenswrapper[4857]: I0222 00:16:54.527841 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9df334b-3beb-41e4-9ac7-cfaaa10ad923" containerName="registry-server" Feb 22 00:16:54 crc kubenswrapper[4857]: E0222 00:16:54.527856 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9df334b-3beb-41e4-9ac7-cfaaa10ad923" containerName="extract-utilities" Feb 22 00:16:54 crc kubenswrapper[4857]: I0222 00:16:54.527863 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9df334b-3beb-41e4-9ac7-cfaaa10ad923" containerName="extract-utilities" Feb 22 00:16:54 crc kubenswrapper[4857]: I0222 00:16:54.527973 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9df334b-3beb-41e4-9ac7-cfaaa10ad923" containerName="registry-server" Feb 22 00:16:54 crc kubenswrapper[4857]: I0222 00:16:54.528801 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q" Feb 22 00:16:54 crc kubenswrapper[4857]: I0222 00:16:54.530407 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 22 00:16:54 crc kubenswrapper[4857]: I0222 00:16:54.535003 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffq8n\" (UniqueName: \"kubernetes.io/projected/489bd4f1-d8e2-4db7-a4b8-76755e099093-kube-api-access-ffq8n\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q\" (UID: \"489bd4f1-d8e2-4db7-a4b8-76755e099093\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q" Feb 22 00:16:54 crc kubenswrapper[4857]: I0222 00:16:54.535090 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/489bd4f1-d8e2-4db7-a4b8-76755e099093-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q\" (UID: \"489bd4f1-d8e2-4db7-a4b8-76755e099093\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q" Feb 22 00:16:54 crc kubenswrapper[4857]: I0222 00:16:54.535130 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/489bd4f1-d8e2-4db7-a4b8-76755e099093-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q\" (UID: \"489bd4f1-d8e2-4db7-a4b8-76755e099093\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q" Feb 22 00:16:54 crc kubenswrapper[4857]: I0222 00:16:54.539363 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q"] Feb 22 00:16:54 crc kubenswrapper[4857]: I0222 00:16:54.636336 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffq8n\" (UniqueName: \"kubernetes.io/projected/489bd4f1-d8e2-4db7-a4b8-76755e099093-kube-api-access-ffq8n\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q\" (UID: \"489bd4f1-d8e2-4db7-a4b8-76755e099093\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q" Feb 22 00:16:54 crc kubenswrapper[4857]: I0222 00:16:54.636374 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/489bd4f1-d8e2-4db7-a4b8-76755e099093-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q\" (UID: \"489bd4f1-d8e2-4db7-a4b8-76755e099093\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q" Feb 22 00:16:54 crc kubenswrapper[4857]: I0222 00:16:54.636397 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/489bd4f1-d8e2-4db7-a4b8-76755e099093-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q\" (UID: \"489bd4f1-d8e2-4db7-a4b8-76755e099093\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q" Feb 22 00:16:54 crc kubenswrapper[4857]: I0222 00:16:54.636841 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/489bd4f1-d8e2-4db7-a4b8-76755e099093-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q\" (UID: \"489bd4f1-d8e2-4db7-a4b8-76755e099093\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q" Feb 22 00:16:54 crc kubenswrapper[4857]: I0222 00:16:54.637019 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/489bd4f1-d8e2-4db7-a4b8-76755e099093-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q\" (UID: \"489bd4f1-d8e2-4db7-a4b8-76755e099093\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q" Feb 22 00:16:54 crc kubenswrapper[4857]: I0222 00:16:54.653669 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffq8n\" (UniqueName: \"kubernetes.io/projected/489bd4f1-d8e2-4db7-a4b8-76755e099093-kube-api-access-ffq8n\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q\" (UID: \"489bd4f1-d8e2-4db7-a4b8-76755e099093\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q" Feb 22 00:16:54 crc kubenswrapper[4857]: I0222 00:16:54.852256 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q" Feb 22 00:16:55 crc kubenswrapper[4857]: I0222 00:16:55.098623 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q"] Feb 22 00:16:55 crc kubenswrapper[4857]: I0222 00:16:55.567188 4857 generic.go:334] "Generic (PLEG): container finished" podID="489bd4f1-d8e2-4db7-a4b8-76755e099093" containerID="3d1d20b6645492d27281eca9a28cc99e6ea7941027a0093fa04ee4cf14241c94" exitCode=0 Feb 22 00:16:55 crc kubenswrapper[4857]: I0222 00:16:55.567237 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q" event={"ID":"489bd4f1-d8e2-4db7-a4b8-76755e099093","Type":"ContainerDied","Data":"3d1d20b6645492d27281eca9a28cc99e6ea7941027a0093fa04ee4cf14241c94"} Feb 22 00:16:55 crc kubenswrapper[4857]: I0222 00:16:55.567265 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q" event={"ID":"489bd4f1-d8e2-4db7-a4b8-76755e099093","Type":"ContainerStarted","Data":"b3cf54f39170aa1989b69afcdb1fb444b75cdd8f6ad385b63ddd24da4bd161ff"} Feb 22 00:16:55 crc kubenswrapper[4857]: I0222 00:16:55.568957 4857 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 22 00:16:57 crc kubenswrapper[4857]: I0222 00:16:57.578739 4857 generic.go:334] "Generic (PLEG): container finished" podID="489bd4f1-d8e2-4db7-a4b8-76755e099093" containerID="5225d7f2b0ea153ff7b61b036bd657c049b282953a1c961bddcd6c876cf51f29" exitCode=0 Feb 22 00:16:57 crc kubenswrapper[4857]: I0222 00:16:57.578786 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q" event={"ID":"489bd4f1-d8e2-4db7-a4b8-76755e099093","Type":"ContainerDied","Data":"5225d7f2b0ea153ff7b61b036bd657c049b282953a1c961bddcd6c876cf51f29"} Feb 22 00:16:58 crc kubenswrapper[4857]: I0222 00:16:58.583997 4857 generic.go:334] "Generic (PLEG): container finished" podID="489bd4f1-d8e2-4db7-a4b8-76755e099093" containerID="b1764f4545bfee1bc82f719adf22b69b8685bdadba678abd19f56b3ba86b64a3" exitCode=0 Feb 22 00:16:58 crc kubenswrapper[4857]: I0222 00:16:58.584713 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q" event={"ID":"489bd4f1-d8e2-4db7-a4b8-76755e099093","Type":"ContainerDied","Data":"b1764f4545bfee1bc82f719adf22b69b8685bdadba678abd19f56b3ba86b64a3"} Feb 22 00:16:59 crc kubenswrapper[4857]: I0222 00:16:59.898422 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q" Feb 22 00:17:00 crc kubenswrapper[4857]: I0222 00:17:00.101199 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/489bd4f1-d8e2-4db7-a4b8-76755e099093-util\") pod \"489bd4f1-d8e2-4db7-a4b8-76755e099093\" (UID: \"489bd4f1-d8e2-4db7-a4b8-76755e099093\") " Feb 22 00:17:00 crc kubenswrapper[4857]: I0222 00:17:00.101824 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/489bd4f1-d8e2-4db7-a4b8-76755e099093-bundle\") pod \"489bd4f1-d8e2-4db7-a4b8-76755e099093\" (UID: \"489bd4f1-d8e2-4db7-a4b8-76755e099093\") " Feb 22 00:17:00 crc kubenswrapper[4857]: I0222 00:17:00.101923 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffq8n\" (UniqueName: \"kubernetes.io/projected/489bd4f1-d8e2-4db7-a4b8-76755e099093-kube-api-access-ffq8n\") pod \"489bd4f1-d8e2-4db7-a4b8-76755e099093\" (UID: \"489bd4f1-d8e2-4db7-a4b8-76755e099093\") " Feb 22 00:17:00 crc kubenswrapper[4857]: I0222 00:17:00.106265 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/489bd4f1-d8e2-4db7-a4b8-76755e099093-bundle" (OuterVolumeSpecName: "bundle") pod "489bd4f1-d8e2-4db7-a4b8-76755e099093" (UID: "489bd4f1-d8e2-4db7-a4b8-76755e099093"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:17:00 crc kubenswrapper[4857]: I0222 00:17:00.109904 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/489bd4f1-d8e2-4db7-a4b8-76755e099093-kube-api-access-ffq8n" (OuterVolumeSpecName: "kube-api-access-ffq8n") pod "489bd4f1-d8e2-4db7-a4b8-76755e099093" (UID: "489bd4f1-d8e2-4db7-a4b8-76755e099093"). InnerVolumeSpecName "kube-api-access-ffq8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:17:00 crc kubenswrapper[4857]: I0222 00:17:00.138687 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/489bd4f1-d8e2-4db7-a4b8-76755e099093-util" (OuterVolumeSpecName: "util") pod "489bd4f1-d8e2-4db7-a4b8-76755e099093" (UID: "489bd4f1-d8e2-4db7-a4b8-76755e099093"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:17:00 crc kubenswrapper[4857]: I0222 00:17:00.203750 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffq8n\" (UniqueName: \"kubernetes.io/projected/489bd4f1-d8e2-4db7-a4b8-76755e099093-kube-api-access-ffq8n\") on node \"crc\" DevicePath \"\"" Feb 22 00:17:00 crc kubenswrapper[4857]: I0222 00:17:00.204004 4857 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/489bd4f1-d8e2-4db7-a4b8-76755e099093-util\") on node \"crc\" DevicePath \"\"" Feb 22 00:17:00 crc kubenswrapper[4857]: I0222 00:17:00.204153 4857 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/489bd4f1-d8e2-4db7-a4b8-76755e099093-bundle\") on node \"crc\" DevicePath \"\"" Feb 22 00:17:00 crc kubenswrapper[4857]: I0222 00:17:00.598317 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q" event={"ID":"489bd4f1-d8e2-4db7-a4b8-76755e099093","Type":"ContainerDied","Data":"b3cf54f39170aa1989b69afcdb1fb444b75cdd8f6ad385b63ddd24da4bd161ff"} Feb 22 00:17:00 crc kubenswrapper[4857]: I0222 00:17:00.598374 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3cf54f39170aa1989b69afcdb1fb444b75cdd8f6ad385b63ddd24da4bd161ff" Feb 22 00:17:00 crc kubenswrapper[4857]: I0222 00:17:00.598377 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q" Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.158380 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6"] Feb 22 00:17:01 crc kubenswrapper[4857]: E0222 00:17:01.158892 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="489bd4f1-d8e2-4db7-a4b8-76755e099093" containerName="pull" Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.158919 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="489bd4f1-d8e2-4db7-a4b8-76755e099093" containerName="pull" Feb 22 00:17:01 crc kubenswrapper[4857]: E0222 00:17:01.158959 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="489bd4f1-d8e2-4db7-a4b8-76755e099093" containerName="util" Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.158975 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="489bd4f1-d8e2-4db7-a4b8-76755e099093" containerName="util" Feb 22 00:17:01 crc kubenswrapper[4857]: E0222 00:17:01.159014 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="489bd4f1-d8e2-4db7-a4b8-76755e099093" containerName="extract" Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.159030 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="489bd4f1-d8e2-4db7-a4b8-76755e099093" containerName="extract" Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.159334 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="489bd4f1-d8e2-4db7-a4b8-76755e099093" containerName="extract" Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.160908 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6" Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.165802 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.174014 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6"] Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.219895 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15ca2571-3005-4eac-80ce-71db200efbfa-util\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6\" (UID: \"15ca2571-3005-4eac-80ce-71db200efbfa\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6" Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.220474 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7thh\" (UniqueName: \"kubernetes.io/projected/15ca2571-3005-4eac-80ce-71db200efbfa-kube-api-access-p7thh\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6\" (UID: \"15ca2571-3005-4eac-80ce-71db200efbfa\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6" Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.220837 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15ca2571-3005-4eac-80ce-71db200efbfa-bundle\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6\" (UID: \"15ca2571-3005-4eac-80ce-71db200efbfa\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6" Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.324111 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15ca2571-3005-4eac-80ce-71db200efbfa-bundle\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6\" (UID: \"15ca2571-3005-4eac-80ce-71db200efbfa\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6" Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.324237 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15ca2571-3005-4eac-80ce-71db200efbfa-util\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6\" (UID: \"15ca2571-3005-4eac-80ce-71db200efbfa\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6" Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.324439 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7thh\" (UniqueName: \"kubernetes.io/projected/15ca2571-3005-4eac-80ce-71db200efbfa-kube-api-access-p7thh\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6\" (UID: \"15ca2571-3005-4eac-80ce-71db200efbfa\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6" Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.325293 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15ca2571-3005-4eac-80ce-71db200efbfa-bundle\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6\" (UID: \"15ca2571-3005-4eac-80ce-71db200efbfa\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6" Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.325837 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15ca2571-3005-4eac-80ce-71db200efbfa-util\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6\" (UID: \"15ca2571-3005-4eac-80ce-71db200efbfa\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6" Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.346898 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7thh\" (UniqueName: \"kubernetes.io/projected/15ca2571-3005-4eac-80ce-71db200efbfa-kube-api-access-p7thh\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6\" (UID: \"15ca2571-3005-4eac-80ce-71db200efbfa\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6" Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.538338 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6" Feb 22 00:17:01 crc kubenswrapper[4857]: I0222 00:17:01.808095 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6"] Feb 22 00:17:02 crc kubenswrapper[4857]: I0222 00:17:02.620311 4857 generic.go:334] "Generic (PLEG): container finished" podID="15ca2571-3005-4eac-80ce-71db200efbfa" containerID="ccf174a74a5fb4ca4dffae05152e2ce94219949af7c66feaabf34f8b9102eaef" exitCode=0 Feb 22 00:17:02 crc kubenswrapper[4857]: I0222 00:17:02.620389 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6" event={"ID":"15ca2571-3005-4eac-80ce-71db200efbfa","Type":"ContainerDied","Data":"ccf174a74a5fb4ca4dffae05152e2ce94219949af7c66feaabf34f8b9102eaef"} Feb 22 00:17:02 crc kubenswrapper[4857]: I0222 00:17:02.620469 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6" event={"ID":"15ca2571-3005-4eac-80ce-71db200efbfa","Type":"ContainerStarted","Data":"cdad187e7bd05a2d2eeafe3f6447b76808587b83c569e6c1344207f640d7a4c7"} Feb 22 00:17:05 crc kubenswrapper[4857]: I0222 00:17:05.650443 4857 generic.go:334] "Generic (PLEG): container finished" podID="15ca2571-3005-4eac-80ce-71db200efbfa" containerID="2c079a4e9cf3a59c5dc3d33eb0343a5138584f6aa279765b86ec176d100cfc69" exitCode=0 Feb 22 00:17:05 crc kubenswrapper[4857]: I0222 00:17:05.650485 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6" event={"ID":"15ca2571-3005-4eac-80ce-71db200efbfa","Type":"ContainerDied","Data":"2c079a4e9cf3a59c5dc3d33eb0343a5138584f6aa279765b86ec176d100cfc69"} Feb 22 00:17:06 crc kubenswrapper[4857]: I0222 00:17:06.657407 4857 generic.go:334] "Generic (PLEG): container finished" podID="15ca2571-3005-4eac-80ce-71db200efbfa" containerID="077ff9d4e2df49c26afd35c43724422cf26855b27e6193c495764cf35fc7b553" exitCode=0 Feb 22 00:17:06 crc kubenswrapper[4857]: I0222 00:17:06.657451 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6" event={"ID":"15ca2571-3005-4eac-80ce-71db200efbfa","Type":"ContainerDied","Data":"077ff9d4e2df49c26afd35c43724422cf26855b27e6193c495764cf35fc7b553"} Feb 22 00:17:07 crc kubenswrapper[4857]: I0222 00:17:07.560856 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8"] Feb 22 00:17:07 crc kubenswrapper[4857]: I0222 00:17:07.562786 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" Feb 22 00:17:07 crc kubenswrapper[4857]: I0222 00:17:07.570137 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8"] Feb 22 00:17:07 crc kubenswrapper[4857]: I0222 00:17:07.600154 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnsqj\" (UniqueName: \"kubernetes.io/projected/8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc-kube-api-access-vnsqj\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8\" (UID: \"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" Feb 22 00:17:07 crc kubenswrapper[4857]: I0222 00:17:07.600304 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8\" (UID: \"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" Feb 22 00:17:07 crc kubenswrapper[4857]: I0222 00:17:07.600338 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8\" (UID: \"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" Feb 22 00:17:07 crc kubenswrapper[4857]: I0222 00:17:07.701736 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnsqj\" (UniqueName: \"kubernetes.io/projected/8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc-kube-api-access-vnsqj\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8\" (UID: \"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" Feb 22 00:17:07 crc kubenswrapper[4857]: I0222 00:17:07.701815 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8\" (UID: \"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" Feb 22 00:17:07 crc kubenswrapper[4857]: I0222 00:17:07.701842 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8\" (UID: \"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" Feb 22 00:17:07 crc kubenswrapper[4857]: I0222 00:17:07.702372 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8\" (UID: \"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" Feb 22 00:17:07 crc kubenswrapper[4857]: I0222 00:17:07.702392 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8\" (UID: \"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" Feb 22 00:17:07 crc kubenswrapper[4857]: I0222 00:17:07.731691 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnsqj\" (UniqueName: \"kubernetes.io/projected/8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc-kube-api-access-vnsqj\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8\" (UID: \"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" Feb 22 00:17:07 crc kubenswrapper[4857]: I0222 00:17:07.878089 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" Feb 22 00:17:07 crc kubenswrapper[4857]: I0222 00:17:07.991474 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6" Feb 22 00:17:08 crc kubenswrapper[4857]: I0222 00:17:08.006601 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15ca2571-3005-4eac-80ce-71db200efbfa-bundle\") pod \"15ca2571-3005-4eac-80ce-71db200efbfa\" (UID: \"15ca2571-3005-4eac-80ce-71db200efbfa\") " Feb 22 00:17:08 crc kubenswrapper[4857]: I0222 00:17:08.006651 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15ca2571-3005-4eac-80ce-71db200efbfa-util\") pod \"15ca2571-3005-4eac-80ce-71db200efbfa\" (UID: \"15ca2571-3005-4eac-80ce-71db200efbfa\") " Feb 22 00:17:08 crc kubenswrapper[4857]: I0222 00:17:08.006700 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7thh\" (UniqueName: \"kubernetes.io/projected/15ca2571-3005-4eac-80ce-71db200efbfa-kube-api-access-p7thh\") pod \"15ca2571-3005-4eac-80ce-71db200efbfa\" (UID: \"15ca2571-3005-4eac-80ce-71db200efbfa\") " Feb 22 00:17:08 crc kubenswrapper[4857]: I0222 00:17:08.007921 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15ca2571-3005-4eac-80ce-71db200efbfa-bundle" (OuterVolumeSpecName: "bundle") pod "15ca2571-3005-4eac-80ce-71db200efbfa" (UID: "15ca2571-3005-4eac-80ce-71db200efbfa"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:17:08 crc kubenswrapper[4857]: I0222 00:17:08.015234 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15ca2571-3005-4eac-80ce-71db200efbfa-kube-api-access-p7thh" (OuterVolumeSpecName: "kube-api-access-p7thh") pod "15ca2571-3005-4eac-80ce-71db200efbfa" (UID: "15ca2571-3005-4eac-80ce-71db200efbfa"). InnerVolumeSpecName "kube-api-access-p7thh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:17:08 crc kubenswrapper[4857]: I0222 00:17:08.034145 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15ca2571-3005-4eac-80ce-71db200efbfa-util" (OuterVolumeSpecName: "util") pod "15ca2571-3005-4eac-80ce-71db200efbfa" (UID: "15ca2571-3005-4eac-80ce-71db200efbfa"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:17:08 crc kubenswrapper[4857]: I0222 00:17:08.107716 4857 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/15ca2571-3005-4eac-80ce-71db200efbfa-bundle\") on node \"crc\" DevicePath \"\"" Feb 22 00:17:08 crc kubenswrapper[4857]: I0222 00:17:08.107745 4857 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/15ca2571-3005-4eac-80ce-71db200efbfa-util\") on node \"crc\" DevicePath \"\"" Feb 22 00:17:08 crc kubenswrapper[4857]: I0222 00:17:08.107754 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7thh\" (UniqueName: \"kubernetes.io/projected/15ca2571-3005-4eac-80ce-71db200efbfa-kube-api-access-p7thh\") on node \"crc\" DevicePath \"\"" Feb 22 00:17:08 crc kubenswrapper[4857]: I0222 00:17:08.289696 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8"] Feb 22 00:17:08 crc kubenswrapper[4857]: I0222 00:17:08.668254 4857 generic.go:334] "Generic (PLEG): container finished" podID="8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc" containerID="a3a0ee45a8840114ac0a22529ee781ef9ddcde118a3789921fc3609e79108d03" exitCode=0 Feb 22 00:17:08 crc kubenswrapper[4857]: I0222 00:17:08.668456 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" event={"ID":"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc","Type":"ContainerDied","Data":"a3a0ee45a8840114ac0a22529ee781ef9ddcde118a3789921fc3609e79108d03"} Feb 22 00:17:08 crc kubenswrapper[4857]: I0222 00:17:08.668565 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" event={"ID":"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc","Type":"ContainerStarted","Data":"ce77d6013435252b980b01ffe26d5c8db5a0c42a7b0826fafecfaa7299614a88"} Feb 22 00:17:08 crc kubenswrapper[4857]: I0222 00:17:08.672434 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6" Feb 22 00:17:08 crc kubenswrapper[4857]: I0222 00:17:08.672918 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6" event={"ID":"15ca2571-3005-4eac-80ce-71db200efbfa","Type":"ContainerDied","Data":"cdad187e7bd05a2d2eeafe3f6447b76808587b83c569e6c1344207f640d7a4c7"} Feb 22 00:17:08 crc kubenswrapper[4857]: I0222 00:17:08.672960 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cdad187e7bd05a2d2eeafe3f6447b76808587b83c569e6c1344207f640d7a4c7" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.161961 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-6zdkm"] Feb 22 00:17:11 crc kubenswrapper[4857]: E0222 00:17:11.162459 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15ca2571-3005-4eac-80ce-71db200efbfa" containerName="pull" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.162471 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="15ca2571-3005-4eac-80ce-71db200efbfa" containerName="pull" Feb 22 00:17:11 crc kubenswrapper[4857]: E0222 00:17:11.162482 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15ca2571-3005-4eac-80ce-71db200efbfa" containerName="extract" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.162487 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="15ca2571-3005-4eac-80ce-71db200efbfa" containerName="extract" Feb 22 00:17:11 crc kubenswrapper[4857]: E0222 00:17:11.162498 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15ca2571-3005-4eac-80ce-71db200efbfa" containerName="util" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.162504 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="15ca2571-3005-4eac-80ce-71db200efbfa" containerName="util" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.162594 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="15ca2571-3005-4eac-80ce-71db200efbfa" containerName="extract" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.162922 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6zdkm" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.164725 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-szcf5" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.168250 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.170056 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.176199 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-6zdkm"] Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.245992 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sbjl\" (UniqueName: \"kubernetes.io/projected/056b8239-b8e8-4610-b38d-8e4cd84ddda5-kube-api-access-7sbjl\") pod \"obo-prometheus-operator-68bc856cb9-6zdkm\" (UID: \"056b8239-b8e8-4610-b38d-8e4cd84ddda5\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6zdkm" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.278181 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx"] Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.278790 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.282286 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.284721 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-vtcxx" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.294446 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx"] Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.302548 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c"] Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.303376 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.347321 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/65eb5aaa-f178-4680-b66b-a468cd347d3a-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c\" (UID: \"65eb5aaa-f178-4680-b66b-a468cd347d3a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.347371 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sbjl\" (UniqueName: \"kubernetes.io/projected/056b8239-b8e8-4610-b38d-8e4cd84ddda5-kube-api-access-7sbjl\") pod \"obo-prometheus-operator-68bc856cb9-6zdkm\" (UID: \"056b8239-b8e8-4610-b38d-8e4cd84ddda5\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6zdkm" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.347396 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/faf26ac1-5a62-47cd-bf57-fe48ffb196cc-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx\" (UID: \"faf26ac1-5a62-47cd-bf57-fe48ffb196cc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.347420 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/faf26ac1-5a62-47cd-bf57-fe48ffb196cc-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx\" (UID: \"faf26ac1-5a62-47cd-bf57-fe48ffb196cc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.347463 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/65eb5aaa-f178-4680-b66b-a468cd347d3a-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c\" (UID: \"65eb5aaa-f178-4680-b66b-a468cd347d3a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.358329 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c"] Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.385233 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sbjl\" (UniqueName: \"kubernetes.io/projected/056b8239-b8e8-4610-b38d-8e4cd84ddda5-kube-api-access-7sbjl\") pod \"obo-prometheus-operator-68bc856cb9-6zdkm\" (UID: \"056b8239-b8e8-4610-b38d-8e4cd84ddda5\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6zdkm" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.448149 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/65eb5aaa-f178-4680-b66b-a468cd347d3a-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c\" (UID: \"65eb5aaa-f178-4680-b66b-a468cd347d3a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.448207 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/faf26ac1-5a62-47cd-bf57-fe48ffb196cc-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx\" (UID: \"faf26ac1-5a62-47cd-bf57-fe48ffb196cc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.448227 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/faf26ac1-5a62-47cd-bf57-fe48ffb196cc-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx\" (UID: \"faf26ac1-5a62-47cd-bf57-fe48ffb196cc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.448248 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/65eb5aaa-f178-4680-b66b-a468cd347d3a-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c\" (UID: \"65eb5aaa-f178-4680-b66b-a468cd347d3a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.452268 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/65eb5aaa-f178-4680-b66b-a468cd347d3a-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c\" (UID: \"65eb5aaa-f178-4680-b66b-a468cd347d3a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.453404 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/faf26ac1-5a62-47cd-bf57-fe48ffb196cc-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx\" (UID: \"faf26ac1-5a62-47cd-bf57-fe48ffb196cc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.455452 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/65eb5aaa-f178-4680-b66b-a468cd347d3a-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c\" (UID: \"65eb5aaa-f178-4680-b66b-a468cd347d3a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.456363 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/faf26ac1-5a62-47cd-bf57-fe48ffb196cc-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx\" (UID: \"faf26ac1-5a62-47cd-bf57-fe48ffb196cc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.482961 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6zdkm" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.487141 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-kcx7r"] Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.487822 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-kcx7r" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.490163 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.490310 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-k27cq" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.502496 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-kcx7r"] Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.550237 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfdnv\" (UniqueName: \"kubernetes.io/projected/bae29f5b-7bae-41c4-a5a0-b5f6e9f57729-kube-api-access-xfdnv\") pod \"observability-operator-59bdc8b94-kcx7r\" (UID: \"bae29f5b-7bae-41c4-a5a0-b5f6e9f57729\") " pod="openshift-operators/observability-operator-59bdc8b94-kcx7r" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.550343 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/bae29f5b-7bae-41c4-a5a0-b5f6e9f57729-observability-operator-tls\") pod \"observability-operator-59bdc8b94-kcx7r\" (UID: \"bae29f5b-7bae-41c4-a5a0-b5f6e9f57729\") " pod="openshift-operators/observability-operator-59bdc8b94-kcx7r" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.595649 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-w8x45"] Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.596359 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-w8x45" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.603780 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-fjhhm" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.617232 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-w8x45"] Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.644521 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.651431 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfdnv\" (UniqueName: \"kubernetes.io/projected/bae29f5b-7bae-41c4-a5a0-b5f6e9f57729-kube-api-access-xfdnv\") pod \"observability-operator-59bdc8b94-kcx7r\" (UID: \"bae29f5b-7bae-41c4-a5a0-b5f6e9f57729\") " pod="openshift-operators/observability-operator-59bdc8b94-kcx7r" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.651502 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c9hw\" (UniqueName: \"kubernetes.io/projected/177b7675-71d3-4c71-8781-4e352edc0044-kube-api-access-2c9hw\") pod \"perses-operator-5bf474d74f-w8x45\" (UID: \"177b7675-71d3-4c71-8781-4e352edc0044\") " pod="openshift-operators/perses-operator-5bf474d74f-w8x45" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.651546 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/177b7675-71d3-4c71-8781-4e352edc0044-openshift-service-ca\") pod \"perses-operator-5bf474d74f-w8x45\" (UID: \"177b7675-71d3-4c71-8781-4e352edc0044\") " pod="openshift-operators/perses-operator-5bf474d74f-w8x45" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.651576 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/bae29f5b-7bae-41c4-a5a0-b5f6e9f57729-observability-operator-tls\") pod \"observability-operator-59bdc8b94-kcx7r\" (UID: \"bae29f5b-7bae-41c4-a5a0-b5f6e9f57729\") " pod="openshift-operators/observability-operator-59bdc8b94-kcx7r" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.665703 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.665996 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/bae29f5b-7bae-41c4-a5a0-b5f6e9f57729-observability-operator-tls\") pod \"observability-operator-59bdc8b94-kcx7r\" (UID: \"bae29f5b-7bae-41c4-a5a0-b5f6e9f57729\") " pod="openshift-operators/observability-operator-59bdc8b94-kcx7r" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.696024 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfdnv\" (UniqueName: \"kubernetes.io/projected/bae29f5b-7bae-41c4-a5a0-b5f6e9f57729-kube-api-access-xfdnv\") pod \"observability-operator-59bdc8b94-kcx7r\" (UID: \"bae29f5b-7bae-41c4-a5a0-b5f6e9f57729\") " pod="openshift-operators/observability-operator-59bdc8b94-kcx7r" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.753080 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c9hw\" (UniqueName: \"kubernetes.io/projected/177b7675-71d3-4c71-8781-4e352edc0044-kube-api-access-2c9hw\") pod \"perses-operator-5bf474d74f-w8x45\" (UID: \"177b7675-71d3-4c71-8781-4e352edc0044\") " pod="openshift-operators/perses-operator-5bf474d74f-w8x45" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.753420 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/177b7675-71d3-4c71-8781-4e352edc0044-openshift-service-ca\") pod \"perses-operator-5bf474d74f-w8x45\" (UID: \"177b7675-71d3-4c71-8781-4e352edc0044\") " pod="openshift-operators/perses-operator-5bf474d74f-w8x45" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.754326 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/177b7675-71d3-4c71-8781-4e352edc0044-openshift-service-ca\") pod \"perses-operator-5bf474d74f-w8x45\" (UID: \"177b7675-71d3-4c71-8781-4e352edc0044\") " pod="openshift-operators/perses-operator-5bf474d74f-w8x45" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.770290 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c9hw\" (UniqueName: \"kubernetes.io/projected/177b7675-71d3-4c71-8781-4e352edc0044-kube-api-access-2c9hw\") pod \"perses-operator-5bf474d74f-w8x45\" (UID: \"177b7675-71d3-4c71-8781-4e352edc0044\") " pod="openshift-operators/perses-operator-5bf474d74f-w8x45" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.816944 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-kcx7r" Feb 22 00:17:11 crc kubenswrapper[4857]: I0222 00:17:11.926799 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-w8x45" Feb 22 00:17:12 crc kubenswrapper[4857]: I0222 00:17:12.703127 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" event={"ID":"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc","Type":"ContainerStarted","Data":"db624ee90091cf98910ac1b3f70d5c527dbfb0b452eee0b92ca86df03cf67053"} Feb 22 00:17:13 crc kubenswrapper[4857]: I0222 00:17:13.017520 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c"] Feb 22 00:17:13 crc kubenswrapper[4857]: W0222 00:17:13.017917 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65eb5aaa_f178_4680_b66b_a468cd347d3a.slice/crio-709ac6fd58a8027cf67cd5c5601a75b4676226807ba45d6b45ccaea0b4ab23b5 WatchSource:0}: Error finding container 709ac6fd58a8027cf67cd5c5601a75b4676226807ba45d6b45ccaea0b4ab23b5: Status 404 returned error can't find the container with id 709ac6fd58a8027cf67cd5c5601a75b4676226807ba45d6b45ccaea0b4ab23b5 Feb 22 00:17:13 crc kubenswrapper[4857]: I0222 00:17:13.030128 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-kcx7r"] Feb 22 00:17:13 crc kubenswrapper[4857]: W0222 00:17:13.044637 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbae29f5b_7bae_41c4_a5a0_b5f6e9f57729.slice/crio-92197b4fc6269456ee5ab2ee2386348ad197c991858f3bd82e140d5714932322 WatchSource:0}: Error finding container 92197b4fc6269456ee5ab2ee2386348ad197c991858f3bd82e140d5714932322: Status 404 returned error can't find the container with id 92197b4fc6269456ee5ab2ee2386348ad197c991858f3bd82e140d5714932322 Feb 22 00:17:13 crc kubenswrapper[4857]: I0222 00:17:13.151680 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-w8x45"] Feb 22 00:17:13 crc kubenswrapper[4857]: W0222 00:17:13.157946 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfaf26ac1_5a62_47cd_bf57_fe48ffb196cc.slice/crio-6fa5dc4b041c9eded401d46148166b6c9ecf7fd699b4c872973222f851ddc354 WatchSource:0}: Error finding container 6fa5dc4b041c9eded401d46148166b6c9ecf7fd699b4c872973222f851ddc354: Status 404 returned error can't find the container with id 6fa5dc4b041c9eded401d46148166b6c9ecf7fd699b4c872973222f851ddc354 Feb 22 00:17:13 crc kubenswrapper[4857]: W0222 00:17:13.161397 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod177b7675_71d3_4c71_8781_4e352edc0044.slice/crio-3e24309459d760c9274af760b3dbe18edf0301c7fece67e62b30b1ad0ba38f04 WatchSource:0}: Error finding container 3e24309459d760c9274af760b3dbe18edf0301c7fece67e62b30b1ad0ba38f04: Status 404 returned error can't find the container with id 3e24309459d760c9274af760b3dbe18edf0301c7fece67e62b30b1ad0ba38f04 Feb 22 00:17:13 crc kubenswrapper[4857]: W0222 00:17:13.164339 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod056b8239_b8e8_4610_b38d_8e4cd84ddda5.slice/crio-d4301c15499c74cca5e4928e1b48cc536584d7de17e47c431e9ad6949cb9b807 WatchSource:0}: Error finding container d4301c15499c74cca5e4928e1b48cc536584d7de17e47c431e9ad6949cb9b807: Status 404 returned error can't find the container with id d4301c15499c74cca5e4928e1b48cc536584d7de17e47c431e9ad6949cb9b807 Feb 22 00:17:13 crc kubenswrapper[4857]: I0222 00:17:13.169538 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx"] Feb 22 00:17:13 crc kubenswrapper[4857]: I0222 00:17:13.174079 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-6zdkm"] Feb 22 00:17:13 crc kubenswrapper[4857]: I0222 00:17:13.711257 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6zdkm" event={"ID":"056b8239-b8e8-4610-b38d-8e4cd84ddda5","Type":"ContainerStarted","Data":"d4301c15499c74cca5e4928e1b48cc536584d7de17e47c431e9ad6949cb9b807"} Feb 22 00:17:13 crc kubenswrapper[4857]: I0222 00:17:13.712229 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx" event={"ID":"faf26ac1-5a62-47cd-bf57-fe48ffb196cc","Type":"ContainerStarted","Data":"6fa5dc4b041c9eded401d46148166b6c9ecf7fd699b4c872973222f851ddc354"} Feb 22 00:17:13 crc kubenswrapper[4857]: I0222 00:17:13.713735 4857 generic.go:334] "Generic (PLEG): container finished" podID="8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc" containerID="db624ee90091cf98910ac1b3f70d5c527dbfb0b452eee0b92ca86df03cf67053" exitCode=0 Feb 22 00:17:13 crc kubenswrapper[4857]: I0222 00:17:13.713821 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" event={"ID":"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc","Type":"ContainerDied","Data":"db624ee90091cf98910ac1b3f70d5c527dbfb0b452eee0b92ca86df03cf67053"} Feb 22 00:17:13 crc kubenswrapper[4857]: I0222 00:17:13.715965 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-kcx7r" event={"ID":"bae29f5b-7bae-41c4-a5a0-b5f6e9f57729","Type":"ContainerStarted","Data":"92197b4fc6269456ee5ab2ee2386348ad197c991858f3bd82e140d5714932322"} Feb 22 00:17:13 crc kubenswrapper[4857]: I0222 00:17:13.718393 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c" event={"ID":"65eb5aaa-f178-4680-b66b-a468cd347d3a","Type":"ContainerStarted","Data":"709ac6fd58a8027cf67cd5c5601a75b4676226807ba45d6b45ccaea0b4ab23b5"} Feb 22 00:17:13 crc kubenswrapper[4857]: I0222 00:17:13.719755 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-w8x45" event={"ID":"177b7675-71d3-4c71-8781-4e352edc0044","Type":"ContainerStarted","Data":"3e24309459d760c9274af760b3dbe18edf0301c7fece67e62b30b1ad0ba38f04"} Feb 22 00:17:14 crc kubenswrapper[4857]: I0222 00:17:14.827766 4857 generic.go:334] "Generic (PLEG): container finished" podID="8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc" containerID="9ba9ee2cb6a4ecbdd3b585d9d9cb1fac9f0939e9ff0219ab40a7520b9fc1ba82" exitCode=0 Feb 22 00:17:14 crc kubenswrapper[4857]: I0222 00:17:14.828174 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" event={"ID":"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc","Type":"ContainerDied","Data":"9ba9ee2cb6a4ecbdd3b585d9d9cb1fac9f0939e9ff0219ab40a7520b9fc1ba82"} Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.643013 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-5449657f9c-6wmjs"] Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.643993 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-5449657f9c-6wmjs" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.646295 4857 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-dockercfg-8m46m" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.646660 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"kube-root-ca.crt" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.647232 4857 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-service-cert" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.647698 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"openshift-service-ca.crt" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.650351 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cb976871-c53b-456a-8cbe-a6a238b7fbb4-webhook-cert\") pod \"elastic-operator-5449657f9c-6wmjs\" (UID: \"cb976871-c53b-456a-8cbe-a6a238b7fbb4\") " pod="service-telemetry/elastic-operator-5449657f9c-6wmjs" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.650407 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cb976871-c53b-456a-8cbe-a6a238b7fbb4-apiservice-cert\") pod \"elastic-operator-5449657f9c-6wmjs\" (UID: \"cb976871-c53b-456a-8cbe-a6a238b7fbb4\") " pod="service-telemetry/elastic-operator-5449657f9c-6wmjs" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.650453 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2rx6\" (UniqueName: \"kubernetes.io/projected/cb976871-c53b-456a-8cbe-a6a238b7fbb4-kube-api-access-d2rx6\") pod \"elastic-operator-5449657f9c-6wmjs\" (UID: \"cb976871-c53b-456a-8cbe-a6a238b7fbb4\") " pod="service-telemetry/elastic-operator-5449657f9c-6wmjs" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.667848 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-5449657f9c-6wmjs"] Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.751773 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cb976871-c53b-456a-8cbe-a6a238b7fbb4-apiservice-cert\") pod \"elastic-operator-5449657f9c-6wmjs\" (UID: \"cb976871-c53b-456a-8cbe-a6a238b7fbb4\") " pod="service-telemetry/elastic-operator-5449657f9c-6wmjs" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.751848 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2rx6\" (UniqueName: \"kubernetes.io/projected/cb976871-c53b-456a-8cbe-a6a238b7fbb4-kube-api-access-d2rx6\") pod \"elastic-operator-5449657f9c-6wmjs\" (UID: \"cb976871-c53b-456a-8cbe-a6a238b7fbb4\") " pod="service-telemetry/elastic-operator-5449657f9c-6wmjs" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.751935 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cb976871-c53b-456a-8cbe-a6a238b7fbb4-webhook-cert\") pod \"elastic-operator-5449657f9c-6wmjs\" (UID: \"cb976871-c53b-456a-8cbe-a6a238b7fbb4\") " pod="service-telemetry/elastic-operator-5449657f9c-6wmjs" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.759664 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cb976871-c53b-456a-8cbe-a6a238b7fbb4-apiservice-cert\") pod \"elastic-operator-5449657f9c-6wmjs\" (UID: \"cb976871-c53b-456a-8cbe-a6a238b7fbb4\") " pod="service-telemetry/elastic-operator-5449657f9c-6wmjs" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.759677 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cb976871-c53b-456a-8cbe-a6a238b7fbb4-webhook-cert\") pod \"elastic-operator-5449657f9c-6wmjs\" (UID: \"cb976871-c53b-456a-8cbe-a6a238b7fbb4\") " pod="service-telemetry/elastic-operator-5449657f9c-6wmjs" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.768543 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2rx6\" (UniqueName: \"kubernetes.io/projected/cb976871-c53b-456a-8cbe-a6a238b7fbb4-kube-api-access-d2rx6\") pod \"elastic-operator-5449657f9c-6wmjs\" (UID: \"cb976871-c53b-456a-8cbe-a6a238b7fbb4\") " pod="service-telemetry/elastic-operator-5449657f9c-6wmjs" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.810121 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.848376 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" event={"ID":"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc","Type":"ContainerDied","Data":"ce77d6013435252b980b01ffe26d5c8db5a0c42a7b0826fafecfaa7299614a88"} Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.848412 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce77d6013435252b980b01ffe26d5c8db5a0c42a7b0826fafecfaa7299614a88" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.848465 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.853064 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc-bundle\") pod \"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc\" (UID: \"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc\") " Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.853175 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc-util\") pod \"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc\" (UID: \"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc\") " Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.853830 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnsqj\" (UniqueName: \"kubernetes.io/projected/8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc-kube-api-access-vnsqj\") pod \"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc\" (UID: \"8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc\") " Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.854156 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc-bundle" (OuterVolumeSpecName: "bundle") pod "8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc" (UID: "8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.867653 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc-util" (OuterVolumeSpecName: "util") pod "8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc" (UID: "8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.872413 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc-kube-api-access-vnsqj" (OuterVolumeSpecName: "kube-api-access-vnsqj") pod "8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc" (UID: "8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc"). InnerVolumeSpecName "kube-api-access-vnsqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.955466 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnsqj\" (UniqueName: \"kubernetes.io/projected/8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc-kube-api-access-vnsqj\") on node \"crc\" DevicePath \"\"" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.955506 4857 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc-bundle\") on node \"crc\" DevicePath \"\"" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.955520 4857 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc-util\") on node \"crc\" DevicePath \"\"" Feb 22 00:17:17 crc kubenswrapper[4857]: I0222 00:17:17.967049 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-5449657f9c-6wmjs" Feb 22 00:17:23 crc kubenswrapper[4857]: I0222 00:17:23.190504 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-5449657f9c-6wmjs"] Feb 22 00:17:23 crc kubenswrapper[4857]: I0222 00:17:23.900327 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6zdkm" event={"ID":"056b8239-b8e8-4610-b38d-8e4cd84ddda5","Type":"ContainerStarted","Data":"9bfdedc6c2c07e2ffa0fc0ad9ee04df4c0a0422f196fa467eeb0f548a98ff7d1"} Feb 22 00:17:23 crc kubenswrapper[4857]: I0222 00:17:23.902006 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-5449657f9c-6wmjs" event={"ID":"cb976871-c53b-456a-8cbe-a6a238b7fbb4","Type":"ContainerStarted","Data":"f5c8cf7063297dc5545df814f5b2f43fba015072c61237ac1ace0127e806d15a"} Feb 22 00:17:23 crc kubenswrapper[4857]: I0222 00:17:23.903119 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx" event={"ID":"faf26ac1-5a62-47cd-bf57-fe48ffb196cc","Type":"ContainerStarted","Data":"ff80aa1108ce7b377812abb1401d1801279a03743c2357de512a0a298fa0a97b"} Feb 22 00:17:23 crc kubenswrapper[4857]: I0222 00:17:23.904101 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-kcx7r" event={"ID":"bae29f5b-7bae-41c4-a5a0-b5f6e9f57729","Type":"ContainerStarted","Data":"47e9a2be711f58c844dd50e22092b695dc464e445b66dd9296d291fb96402f3a"} Feb 22 00:17:23 crc kubenswrapper[4857]: I0222 00:17:23.904218 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-kcx7r" Feb 22 00:17:23 crc kubenswrapper[4857]: I0222 00:17:23.905422 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c" event={"ID":"65eb5aaa-f178-4680-b66b-a468cd347d3a","Type":"ContainerStarted","Data":"a03143fe1573df10b5c5e820dd487e69db16949aa978a7abd9f14d3a04c698e1"} Feb 22 00:17:23 crc kubenswrapper[4857]: I0222 00:17:23.906636 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-w8x45" event={"ID":"177b7675-71d3-4c71-8781-4e352edc0044","Type":"ContainerStarted","Data":"1fcfdd39bf4869de32ed25bc60ff7d219cc82e9c9f1a44e6fbc25482cba47db5"} Feb 22 00:17:23 crc kubenswrapper[4857]: I0222 00:17:23.906762 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-w8x45" Feb 22 00:17:23 crc kubenswrapper[4857]: I0222 00:17:23.921144 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-6zdkm" podStartSLOduration=3.077317981 podStartE2EDuration="12.92112569s" podCreationTimestamp="2026-02-22 00:17:11 +0000 UTC" firstStartedPulling="2026-02-22 00:17:13.168249549 +0000 UTC m=+640.806978802" lastFinishedPulling="2026-02-22 00:17:23.012057258 +0000 UTC m=+650.650786511" observedRunningTime="2026-02-22 00:17:23.9171867 +0000 UTC m=+651.555915953" watchObservedRunningTime="2026-02-22 00:17:23.92112569 +0000 UTC m=+651.559854933" Feb 22 00:17:23 crc kubenswrapper[4857]: I0222 00:17:23.943832 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-kcx7r" podStartSLOduration=2.989000783 podStartE2EDuration="12.943813628s" podCreationTimestamp="2026-02-22 00:17:11 +0000 UTC" firstStartedPulling="2026-02-22 00:17:13.046632631 +0000 UTC m=+640.685361884" lastFinishedPulling="2026-02-22 00:17:23.001445476 +0000 UTC m=+650.640174729" observedRunningTime="2026-02-22 00:17:23.938667011 +0000 UTC m=+651.577396264" watchObservedRunningTime="2026-02-22 00:17:23.943813628 +0000 UTC m=+651.582542881" Feb 22 00:17:23 crc kubenswrapper[4857]: I0222 00:17:23.966913 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-kcx7r" Feb 22 00:17:23 crc kubenswrapper[4857]: I0222 00:17:23.972977 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c" podStartSLOduration=3.007559554 podStartE2EDuration="12.972959112s" podCreationTimestamp="2026-02-22 00:17:11 +0000 UTC" firstStartedPulling="2026-02-22 00:17:13.02029909 +0000 UTC m=+640.659028343" lastFinishedPulling="2026-02-22 00:17:22.985698648 +0000 UTC m=+650.624427901" observedRunningTime="2026-02-22 00:17:23.969691773 +0000 UTC m=+651.608421026" watchObservedRunningTime="2026-02-22 00:17:23.972959112 +0000 UTC m=+651.611688365" Feb 22 00:17:24 crc kubenswrapper[4857]: I0222 00:17:24.020585 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-w8x45" podStartSLOduration=3.172898298 podStartE2EDuration="13.020569156s" podCreationTimestamp="2026-02-22 00:17:11 +0000 UTC" firstStartedPulling="2026-02-22 00:17:13.165735411 +0000 UTC m=+640.804464664" lastFinishedPulling="2026-02-22 00:17:23.013406269 +0000 UTC m=+650.652135522" observedRunningTime="2026-02-22 00:17:23.995677661 +0000 UTC m=+651.634406914" watchObservedRunningTime="2026-02-22 00:17:24.020569156 +0000 UTC m=+651.659298409" Feb 22 00:17:24 crc kubenswrapper[4857]: I0222 00:17:24.051377 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx" podStartSLOduration=3.196255812 podStartE2EDuration="13.051363611s" podCreationTimestamp="2026-02-22 00:17:11 +0000 UTC" firstStartedPulling="2026-02-22 00:17:13.161280851 +0000 UTC m=+640.800010104" lastFinishedPulling="2026-02-22 00:17:23.01638865 +0000 UTC m=+650.655117903" observedRunningTime="2026-02-22 00:17:24.023605069 +0000 UTC m=+651.662334322" watchObservedRunningTime="2026-02-22 00:17:24.051363611 +0000 UTC m=+651.690092864" Feb 22 00:17:26 crc kubenswrapper[4857]: I0222 00:17:26.922655 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-5449657f9c-6wmjs" event={"ID":"cb976871-c53b-456a-8cbe-a6a238b7fbb4","Type":"ContainerStarted","Data":"239dd220785c3141ea10693b2a3fda10402e80ef122f2cadc76c8339bf5b7d13"} Feb 22 00:17:26 crc kubenswrapper[4857]: I0222 00:17:26.953641 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-5449657f9c-6wmjs" podStartSLOduration=7.198588418 podStartE2EDuration="9.953624546s" podCreationTimestamp="2026-02-22 00:17:17 +0000 UTC" firstStartedPulling="2026-02-22 00:17:23.20361976 +0000 UTC m=+650.842349013" lastFinishedPulling="2026-02-22 00:17:25.958655868 +0000 UTC m=+653.597385141" observedRunningTime="2026-02-22 00:17:26.950333156 +0000 UTC m=+654.589062399" watchObservedRunningTime="2026-02-22 00:17:26.953624546 +0000 UTC m=+654.592353799" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.357151 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 22 00:17:30 crc kubenswrapper[4857]: E0222 00:17:30.357845 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc" containerName="util" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.357855 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc" containerName="util" Feb 22 00:17:30 crc kubenswrapper[4857]: E0222 00:17:30.357873 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc" containerName="pull" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.357880 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc" containerName="pull" Feb 22 00:17:30 crc kubenswrapper[4857]: E0222 00:17:30.357888 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc" containerName="extract" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.357894 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc" containerName="extract" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.357981 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc" containerName="extract" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.358689 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.361578 4857 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-config" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.361805 4857 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-transport-certs" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.362271 4857 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-internal-users" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.362646 4857 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-remote-ca" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.362802 4857 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-dockercfg-hw6nd" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.362947 4857 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-http-certs-internal" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.363172 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-scripts" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.363330 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-unicast-hosts" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.363825 4857 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-xpack-file-realm" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.380699 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.531701 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.531766 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.531804 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.531834 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.531861 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/ee783ac5-b162-4821-87ff-09dbc60e98f1-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.531888 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/ee783ac5-b162-4821-87ff-09dbc60e98f1-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.531911 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.531941 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.531964 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/ee783ac5-b162-4821-87ff-09dbc60e98f1-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.531996 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.532020 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.532089 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.532112 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.532138 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/ee783ac5-b162-4821-87ff-09dbc60e98f1-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.532160 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.633671 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.633711 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.633734 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/ee783ac5-b162-4821-87ff-09dbc60e98f1-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.633755 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.633809 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.633836 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.633859 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.633879 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.633900 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/ee783ac5-b162-4821-87ff-09dbc60e98f1-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.633918 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/ee783ac5-b162-4821-87ff-09dbc60e98f1-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.633933 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.633952 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.633969 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/ee783ac5-b162-4821-87ff-09dbc60e98f1-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.633990 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.634005 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.634816 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/ee783ac5-b162-4821-87ff-09dbc60e98f1-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.634846 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.635156 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/ee783ac5-b162-4821-87ff-09dbc60e98f1-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.635210 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/ee783ac5-b162-4821-87ff-09dbc60e98f1-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.635744 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.635948 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.636334 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.636546 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.639995 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.640131 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.640135 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.640166 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/ee783ac5-b162-4821-87ff-09dbc60e98f1-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.640799 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.640856 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.642109 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/ee783ac5-b162-4821-87ff-09dbc60e98f1-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"ee783ac5-b162-4821-87ff-09dbc60e98f1\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:30 crc kubenswrapper[4857]: I0222 00:17:30.679574 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:31 crc kubenswrapper[4857]: I0222 00:17:31.158174 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 22 00:17:31 crc kubenswrapper[4857]: I0222 00:17:31.932668 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-w8x45" Feb 22 00:17:31 crc kubenswrapper[4857]: I0222 00:17:31.945894 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"ee783ac5-b162-4821-87ff-09dbc60e98f1","Type":"ContainerStarted","Data":"33f9574bd60391ca20a1b13f400d567e24ade0069fa57e513e03ad4a0cb79d53"} Feb 22 00:17:33 crc kubenswrapper[4857]: I0222 00:17:33.283932 4857 scope.go:117] "RemoveContainer" containerID="6f772d5bf4afd3597656f75b78ddd28f1cc5111bfbe4e0056153e48b09206633" Feb 22 00:17:33 crc kubenswrapper[4857]: I0222 00:17:33.620502 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-b6p26"] Feb 22 00:17:33 crc kubenswrapper[4857]: I0222 00:17:33.621265 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-b6p26" Feb 22 00:17:33 crc kubenswrapper[4857]: I0222 00:17:33.623852 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Feb 22 00:17:33 crc kubenswrapper[4857]: I0222 00:17:33.624096 4857 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-tkwpm" Feb 22 00:17:33 crc kubenswrapper[4857]: I0222 00:17:33.624416 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Feb 22 00:17:33 crc kubenswrapper[4857]: I0222 00:17:33.668277 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/457afdd3-7737-481f-a527-e1df18a67852-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-b6p26\" (UID: \"457afdd3-7737-481f-a527-e1df18a67852\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-b6p26" Feb 22 00:17:33 crc kubenswrapper[4857]: I0222 00:17:33.668371 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rj9w\" (UniqueName: \"kubernetes.io/projected/457afdd3-7737-481f-a527-e1df18a67852-kube-api-access-6rj9w\") pod \"cert-manager-operator-controller-manager-5586865c96-b6p26\" (UID: \"457afdd3-7737-481f-a527-e1df18a67852\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-b6p26" Feb 22 00:17:33 crc kubenswrapper[4857]: I0222 00:17:33.703821 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-b6p26"] Feb 22 00:17:33 crc kubenswrapper[4857]: I0222 00:17:33.769273 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rj9w\" (UniqueName: \"kubernetes.io/projected/457afdd3-7737-481f-a527-e1df18a67852-kube-api-access-6rj9w\") pod \"cert-manager-operator-controller-manager-5586865c96-b6p26\" (UID: \"457afdd3-7737-481f-a527-e1df18a67852\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-b6p26" Feb 22 00:17:33 crc kubenswrapper[4857]: I0222 00:17:33.769328 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/457afdd3-7737-481f-a527-e1df18a67852-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-b6p26\" (UID: \"457afdd3-7737-481f-a527-e1df18a67852\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-b6p26" Feb 22 00:17:33 crc kubenswrapper[4857]: I0222 00:17:33.769852 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/457afdd3-7737-481f-a527-e1df18a67852-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-b6p26\" (UID: \"457afdd3-7737-481f-a527-e1df18a67852\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-b6p26" Feb 22 00:17:33 crc kubenswrapper[4857]: I0222 00:17:33.788845 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rj9w\" (UniqueName: \"kubernetes.io/projected/457afdd3-7737-481f-a527-e1df18a67852-kube-api-access-6rj9w\") pod \"cert-manager-operator-controller-manager-5586865c96-b6p26\" (UID: \"457afdd3-7737-481f-a527-e1df18a67852\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-b6p26" Feb 22 00:17:33 crc kubenswrapper[4857]: I0222 00:17:33.936865 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-b6p26" Feb 22 00:17:34 crc kubenswrapper[4857]: I0222 00:17:34.256082 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-b6p26"] Feb 22 00:17:34 crc kubenswrapper[4857]: W0222 00:17:34.271282 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod457afdd3_7737_481f_a527_e1df18a67852.slice/crio-17efc998fece249aa73f51e6de345faa00ebdfde71836176acbf4fdc5e39e8cb WatchSource:0}: Error finding container 17efc998fece249aa73f51e6de345faa00ebdfde71836176acbf4fdc5e39e8cb: Status 404 returned error can't find the container with id 17efc998fece249aa73f51e6de345faa00ebdfde71836176acbf4fdc5e39e8cb Feb 22 00:17:34 crc kubenswrapper[4857]: I0222 00:17:34.967254 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-b6p26" event={"ID":"457afdd3-7737-481f-a527-e1df18a67852","Type":"ContainerStarted","Data":"17efc998fece249aa73f51e6de345faa00ebdfde71836176acbf4fdc5e39e8cb"} Feb 22 00:17:42 crc kubenswrapper[4857]: I0222 00:17:42.027660 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-b6p26" event={"ID":"457afdd3-7737-481f-a527-e1df18a67852","Type":"ContainerStarted","Data":"020cbc54b5ebce2610d0fda5a560b71adadc12f69cf8a1ad27b6897ce089c585"} Feb 22 00:17:42 crc kubenswrapper[4857]: I0222 00:17:42.047779 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-b6p26" podStartSLOduration=1.559094545 podStartE2EDuration="9.047763243s" podCreationTimestamp="2026-02-22 00:17:33 +0000 UTC" firstStartedPulling="2026-02-22 00:17:34.277518292 +0000 UTC m=+661.916247545" lastFinishedPulling="2026-02-22 00:17:41.76618699 +0000 UTC m=+669.404916243" observedRunningTime="2026-02-22 00:17:42.046056461 +0000 UTC m=+669.684785714" watchObservedRunningTime="2026-02-22 00:17:42.047763243 +0000 UTC m=+669.686492496" Feb 22 00:17:45 crc kubenswrapper[4857]: I0222 00:17:45.318440 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-qphbj"] Feb 22 00:17:45 crc kubenswrapper[4857]: I0222 00:17:45.319456 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-qphbj" Feb 22 00:17:45 crc kubenswrapper[4857]: I0222 00:17:45.321111 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 22 00:17:45 crc kubenswrapper[4857]: I0222 00:17:45.323877 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 22 00:17:45 crc kubenswrapper[4857]: I0222 00:17:45.324082 4857 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-g4b2c" Feb 22 00:17:45 crc kubenswrapper[4857]: I0222 00:17:45.332373 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-qphbj"] Feb 22 00:17:45 crc kubenswrapper[4857]: I0222 00:17:45.351487 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f9b6b93d-c3b5-4165-a9a6-28274a241b7a-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-qphbj\" (UID: \"f9b6b93d-c3b5-4165-a9a6-28274a241b7a\") " pod="cert-manager/cert-manager-webhook-6888856db4-qphbj" Feb 22 00:17:45 crc kubenswrapper[4857]: I0222 00:17:45.351574 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84hps\" (UniqueName: \"kubernetes.io/projected/f9b6b93d-c3b5-4165-a9a6-28274a241b7a-kube-api-access-84hps\") pod \"cert-manager-webhook-6888856db4-qphbj\" (UID: \"f9b6b93d-c3b5-4165-a9a6-28274a241b7a\") " pod="cert-manager/cert-manager-webhook-6888856db4-qphbj" Feb 22 00:17:45 crc kubenswrapper[4857]: I0222 00:17:45.452453 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f9b6b93d-c3b5-4165-a9a6-28274a241b7a-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-qphbj\" (UID: \"f9b6b93d-c3b5-4165-a9a6-28274a241b7a\") " pod="cert-manager/cert-manager-webhook-6888856db4-qphbj" Feb 22 00:17:45 crc kubenswrapper[4857]: I0222 00:17:45.452527 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84hps\" (UniqueName: \"kubernetes.io/projected/f9b6b93d-c3b5-4165-a9a6-28274a241b7a-kube-api-access-84hps\") pod \"cert-manager-webhook-6888856db4-qphbj\" (UID: \"f9b6b93d-c3b5-4165-a9a6-28274a241b7a\") " pod="cert-manager/cert-manager-webhook-6888856db4-qphbj" Feb 22 00:17:45 crc kubenswrapper[4857]: I0222 00:17:45.473791 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f9b6b93d-c3b5-4165-a9a6-28274a241b7a-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-qphbj\" (UID: \"f9b6b93d-c3b5-4165-a9a6-28274a241b7a\") " pod="cert-manager/cert-manager-webhook-6888856db4-qphbj" Feb 22 00:17:45 crc kubenswrapper[4857]: I0222 00:17:45.482839 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84hps\" (UniqueName: \"kubernetes.io/projected/f9b6b93d-c3b5-4165-a9a6-28274a241b7a-kube-api-access-84hps\") pod \"cert-manager-webhook-6888856db4-qphbj\" (UID: \"f9b6b93d-c3b5-4165-a9a6-28274a241b7a\") " pod="cert-manager/cert-manager-webhook-6888856db4-qphbj" Feb 22 00:17:45 crc kubenswrapper[4857]: I0222 00:17:45.645843 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-qphbj" Feb 22 00:17:46 crc kubenswrapper[4857]: W0222 00:17:46.068825 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9b6b93d_c3b5_4165_a9a6_28274a241b7a.slice/crio-0f55457f20c270c23f5ea5174c29e23dda59c437b255b946e758b5d960133cda WatchSource:0}: Error finding container 0f55457f20c270c23f5ea5174c29e23dda59c437b255b946e758b5d960133cda: Status 404 returned error can't find the container with id 0f55457f20c270c23f5ea5174c29e23dda59c437b255b946e758b5d960133cda Feb 22 00:17:46 crc kubenswrapper[4857]: I0222 00:17:46.071385 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-qphbj"] Feb 22 00:17:47 crc kubenswrapper[4857]: I0222 00:17:47.055066 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-qphbj" event={"ID":"f9b6b93d-c3b5-4165-a9a6-28274a241b7a","Type":"ContainerStarted","Data":"0f55457f20c270c23f5ea5174c29e23dda59c437b255b946e758b5d960133cda"} Feb 22 00:17:48 crc kubenswrapper[4857]: I0222 00:17:48.809143 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-gpl5d"] Feb 22 00:17:48 crc kubenswrapper[4857]: I0222 00:17:48.811967 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-gpl5d" Feb 22 00:17:48 crc kubenswrapper[4857]: I0222 00:17:48.816386 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-gpl5d"] Feb 22 00:17:48 crc kubenswrapper[4857]: I0222 00:17:48.817272 4857 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-c98dl" Feb 22 00:17:48 crc kubenswrapper[4857]: I0222 00:17:48.897919 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d68808f6-f5f0-452f-9036-abce9664d3e7-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-gpl5d\" (UID: \"d68808f6-f5f0-452f-9036-abce9664d3e7\") " pod="cert-manager/cert-manager-cainjector-5545bd876-gpl5d" Feb 22 00:17:48 crc kubenswrapper[4857]: I0222 00:17:48.897987 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq5qh\" (UniqueName: \"kubernetes.io/projected/d68808f6-f5f0-452f-9036-abce9664d3e7-kube-api-access-tq5qh\") pod \"cert-manager-cainjector-5545bd876-gpl5d\" (UID: \"d68808f6-f5f0-452f-9036-abce9664d3e7\") " pod="cert-manager/cert-manager-cainjector-5545bd876-gpl5d" Feb 22 00:17:49 crc kubenswrapper[4857]: I0222 00:17:49.008874 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d68808f6-f5f0-452f-9036-abce9664d3e7-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-gpl5d\" (UID: \"d68808f6-f5f0-452f-9036-abce9664d3e7\") " pod="cert-manager/cert-manager-cainjector-5545bd876-gpl5d" Feb 22 00:17:49 crc kubenswrapper[4857]: I0222 00:17:49.008974 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq5qh\" (UniqueName: \"kubernetes.io/projected/d68808f6-f5f0-452f-9036-abce9664d3e7-kube-api-access-tq5qh\") pod \"cert-manager-cainjector-5545bd876-gpl5d\" (UID: \"d68808f6-f5f0-452f-9036-abce9664d3e7\") " pod="cert-manager/cert-manager-cainjector-5545bd876-gpl5d" Feb 22 00:17:49 crc kubenswrapper[4857]: I0222 00:17:49.030875 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d68808f6-f5f0-452f-9036-abce9664d3e7-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-gpl5d\" (UID: \"d68808f6-f5f0-452f-9036-abce9664d3e7\") " pod="cert-manager/cert-manager-cainjector-5545bd876-gpl5d" Feb 22 00:17:49 crc kubenswrapper[4857]: I0222 00:17:49.041532 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq5qh\" (UniqueName: \"kubernetes.io/projected/d68808f6-f5f0-452f-9036-abce9664d3e7-kube-api-access-tq5qh\") pod \"cert-manager-cainjector-5545bd876-gpl5d\" (UID: \"d68808f6-f5f0-452f-9036-abce9664d3e7\") " pod="cert-manager/cert-manager-cainjector-5545bd876-gpl5d" Feb 22 00:17:49 crc kubenswrapper[4857]: I0222 00:17:49.130431 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-gpl5d" Feb 22 00:17:51 crc kubenswrapper[4857]: I0222 00:17:51.406511 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-gpl5d"] Feb 22 00:17:52 crc kubenswrapper[4857]: I0222 00:17:52.082199 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-gpl5d" event={"ID":"d68808f6-f5f0-452f-9036-abce9664d3e7","Type":"ContainerStarted","Data":"c735458356229d594a35561f434548400d9fdf2279556d0cebd4f8a81d5235b5"} Feb 22 00:17:52 crc kubenswrapper[4857]: I0222 00:17:52.083728 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"ee783ac5-b162-4821-87ff-09dbc60e98f1","Type":"ContainerStarted","Data":"9de8247c10301d48e616a225c2a2364692a4c75d8abdb132f9edcc9fcf697c3e"} Feb 22 00:17:52 crc kubenswrapper[4857]: I0222 00:17:52.251992 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 22 00:17:52 crc kubenswrapper[4857]: I0222 00:17:52.283773 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 22 00:17:53 crc kubenswrapper[4857]: I0222 00:17:53.094362 4857 generic.go:334] "Generic (PLEG): container finished" podID="ee783ac5-b162-4821-87ff-09dbc60e98f1" containerID="9de8247c10301d48e616a225c2a2364692a4c75d8abdb132f9edcc9fcf697c3e" exitCode=0 Feb 22 00:17:53 crc kubenswrapper[4857]: I0222 00:17:53.094462 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"ee783ac5-b162-4821-87ff-09dbc60e98f1","Type":"ContainerDied","Data":"9de8247c10301d48e616a225c2a2364692a4c75d8abdb132f9edcc9fcf697c3e"} Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.454302 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-framework-index-1-build"] Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.460893 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.464719 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-1-ca" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.464856 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-1-global-ca" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.464876 4857 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-l2p5g" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.465024 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-1-sys-config" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.465320 4857 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"service-telemetry-framework-index-dockercfg" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.480350 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-container-storage-run\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.480395 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5707c235-63be-4f94-8feb-1ae7c83b7926-buildcachedir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.480413 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/5707c235-63be-4f94-8feb-1ae7c83b7926-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.480437 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5707c235-63be-4f94-8feb-1ae7c83b7926-build-system-configs\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.480479 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-buildworkdir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.480506 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-container-storage-root\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.480541 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/5707c235-63be-4f94-8feb-1ae7c83b7926-builder-dockercfg-l2p5g-pull\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.480558 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5707c235-63be-4f94-8feb-1ae7c83b7926-build-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.480581 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xl9r\" (UniqueName: \"kubernetes.io/projected/5707c235-63be-4f94-8feb-1ae7c83b7926-kube-api-access-6xl9r\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.480599 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5707c235-63be-4f94-8feb-1ae7c83b7926-node-pullsecrets\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.480620 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/5707c235-63be-4f94-8feb-1ae7c83b7926-builder-dockercfg-l2p5g-push\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.480650 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5707c235-63be-4f94-8feb-1ae7c83b7926-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.480665 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-build-blob-cache\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.484579 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-framework-index-1-build"] Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.582156 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/5707c235-63be-4f94-8feb-1ae7c83b7926-builder-dockercfg-l2p5g-push\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.582215 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5707c235-63be-4f94-8feb-1ae7c83b7926-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.582236 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-build-blob-cache\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.582262 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-container-storage-run\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.582284 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5707c235-63be-4f94-8feb-1ae7c83b7926-buildcachedir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.582302 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/5707c235-63be-4f94-8feb-1ae7c83b7926-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.582326 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5707c235-63be-4f94-8feb-1ae7c83b7926-build-system-configs\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.582347 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-buildworkdir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.582369 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-container-storage-root\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.582401 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/5707c235-63be-4f94-8feb-1ae7c83b7926-builder-dockercfg-l2p5g-pull\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.582406 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5707c235-63be-4f94-8feb-1ae7c83b7926-buildcachedir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.582420 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5707c235-63be-4f94-8feb-1ae7c83b7926-build-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.582527 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xl9r\" (UniqueName: \"kubernetes.io/projected/5707c235-63be-4f94-8feb-1ae7c83b7926-kube-api-access-6xl9r\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.582568 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5707c235-63be-4f94-8feb-1ae7c83b7926-node-pullsecrets\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.582766 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5707c235-63be-4f94-8feb-1ae7c83b7926-node-pullsecrets\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.582830 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-build-blob-cache\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.582979 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-container-storage-run\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.583102 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-buildworkdir\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.583160 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5707c235-63be-4f94-8feb-1ae7c83b7926-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.583343 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-container-storage-root\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.583361 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5707c235-63be-4f94-8feb-1ae7c83b7926-build-ca-bundles\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.583571 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5707c235-63be-4f94-8feb-1ae7c83b7926-build-system-configs\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.588028 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/5707c235-63be-4f94-8feb-1ae7c83b7926-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.588066 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/5707c235-63be-4f94-8feb-1ae7c83b7926-builder-dockercfg-l2p5g-pull\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.594004 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/5707c235-63be-4f94-8feb-1ae7c83b7926-builder-dockercfg-l2p5g-push\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.604458 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xl9r\" (UniqueName: \"kubernetes.io/projected/5707c235-63be-4f94-8feb-1ae7c83b7926-kube-api-access-6xl9r\") pod \"service-telemetry-framework-index-1-build\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:54 crc kubenswrapper[4857]: I0222 00:17:54.847113 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:17:55 crc kubenswrapper[4857]: I0222 00:17:55.056065 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-framework-index-1-build"] Feb 22 00:17:55 crc kubenswrapper[4857]: I0222 00:17:55.113641 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-qphbj" event={"ID":"f9b6b93d-c3b5-4165-a9a6-28274a241b7a","Type":"ContainerStarted","Data":"4a0368e33b95f5781048d02b0ea65e1943be0537f49765803e33f2cced9ace09"} Feb 22 00:17:55 crc kubenswrapper[4857]: I0222 00:17:55.113780 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-6888856db4-qphbj" Feb 22 00:17:55 crc kubenswrapper[4857]: I0222 00:17:55.114816 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"5707c235-63be-4f94-8feb-1ae7c83b7926","Type":"ContainerStarted","Data":"f1659231b2b5e5f2b1df70da94b8531eceef08d644d709bd886cb494b4356edc"} Feb 22 00:17:55 crc kubenswrapper[4857]: I0222 00:17:55.116462 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-gpl5d" event={"ID":"d68808f6-f5f0-452f-9036-abce9664d3e7","Type":"ContainerStarted","Data":"e96b38a5527eba330e752d3fa51581b3f1a03e0aceac01d10dd1543bf3bbea66"} Feb 22 00:17:55 crc kubenswrapper[4857]: I0222 00:17:55.118978 4857 generic.go:334] "Generic (PLEG): container finished" podID="ee783ac5-b162-4821-87ff-09dbc60e98f1" containerID="8e2aaf5c3f170d4ed02d6d5d86345eaf1f891d40b7d010a2f9e148c8c927c716" exitCode=0 Feb 22 00:17:55 crc kubenswrapper[4857]: I0222 00:17:55.119019 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"ee783ac5-b162-4821-87ff-09dbc60e98f1","Type":"ContainerDied","Data":"8e2aaf5c3f170d4ed02d6d5d86345eaf1f891d40b7d010a2f9e148c8c927c716"} Feb 22 00:17:55 crc kubenswrapper[4857]: I0222 00:17:55.133836 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-6888856db4-qphbj" podStartSLOduration=1.926125812 podStartE2EDuration="10.132427742s" podCreationTimestamp="2026-02-22 00:17:45 +0000 UTC" firstStartedPulling="2026-02-22 00:17:46.070898755 +0000 UTC m=+673.709628008" lastFinishedPulling="2026-02-22 00:17:54.277200685 +0000 UTC m=+681.915929938" observedRunningTime="2026-02-22 00:17:55.12774857 +0000 UTC m=+682.766477833" watchObservedRunningTime="2026-02-22 00:17:55.132427742 +0000 UTC m=+682.771157015" Feb 22 00:17:55 crc kubenswrapper[4857]: I0222 00:17:55.191647 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-5545bd876-gpl5d" podStartSLOduration=4.323814858 podStartE2EDuration="7.191631788s" podCreationTimestamp="2026-02-22 00:17:48 +0000 UTC" firstStartedPulling="2026-02-22 00:17:51.427660229 +0000 UTC m=+679.066389482" lastFinishedPulling="2026-02-22 00:17:54.295477159 +0000 UTC m=+681.934206412" observedRunningTime="2026-02-22 00:17:55.144265511 +0000 UTC m=+682.782994764" watchObservedRunningTime="2026-02-22 00:17:55.191631788 +0000 UTC m=+682.830361041" Feb 22 00:17:56 crc kubenswrapper[4857]: I0222 00:17:56.127522 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"ee783ac5-b162-4821-87ff-09dbc60e98f1","Type":"ContainerStarted","Data":"90a1573e50aa043a4ef95946b0d35b353974f7ead6060d6c9f568bce4eef33c4"} Feb 22 00:17:56 crc kubenswrapper[4857]: I0222 00:17:56.129230 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:17:56 crc kubenswrapper[4857]: I0222 00:17:56.161834 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=6.062231019 podStartE2EDuration="26.161809043s" podCreationTimestamp="2026-02-22 00:17:30 +0000 UTC" firstStartedPulling="2026-02-22 00:17:31.165285817 +0000 UTC m=+658.804015070" lastFinishedPulling="2026-02-22 00:17:51.264863841 +0000 UTC m=+678.903593094" observedRunningTime="2026-02-22 00:17:56.156994377 +0000 UTC m=+683.795723630" watchObservedRunningTime="2026-02-22 00:17:56.161809043 +0000 UTC m=+683.800538286" Feb 22 00:18:00 crc kubenswrapper[4857]: I0222 00:18:00.648524 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-6888856db4-qphbj" Feb 22 00:18:01 crc kubenswrapper[4857]: I0222 00:18:01.156334 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"5707c235-63be-4f94-8feb-1ae7c83b7926","Type":"ContainerStarted","Data":"ac05211744cab7f7ce160169a908907ddb3fb0344b4ec27214992fd339e9f8a5"} Feb 22 00:18:01 crc kubenswrapper[4857]: E0222 00:18:01.224216 4857 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=8528648506435008715, SKID=, AKID=49:B4:28:90:5C:47:E0:09:27:27:EB:F1:16:CB:58:04:C2:1D:C0:A3 failed: x509: certificate signed by unknown authority" Feb 22 00:18:02 crc kubenswrapper[4857]: I0222 00:18:02.247606 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-framework-index-1-build"] Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.166470 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-framework-index-1-build" podUID="5707c235-63be-4f94-8feb-1ae7c83b7926" containerName="git-clone" containerID="cri-o://ac05211744cab7f7ce160169a908907ddb3fb0344b4ec27214992fd339e9f8a5" gracePeriod=30 Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.518322 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-framework-index-1-build_5707c235-63be-4f94-8feb-1ae7c83b7926/git-clone/0.log" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.518720 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.597428 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5707c235-63be-4f94-8feb-1ae7c83b7926-build-proxy-ca-bundles\") pod \"5707c235-63be-4f94-8feb-1ae7c83b7926\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.597483 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-container-storage-run\") pod \"5707c235-63be-4f94-8feb-1ae7c83b7926\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.597519 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xl9r\" (UniqueName: \"kubernetes.io/projected/5707c235-63be-4f94-8feb-1ae7c83b7926-kube-api-access-6xl9r\") pod \"5707c235-63be-4f94-8feb-1ae7c83b7926\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.597546 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/5707c235-63be-4f94-8feb-1ae7c83b7926-builder-dockercfg-l2p5g-pull\") pod \"5707c235-63be-4f94-8feb-1ae7c83b7926\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.597578 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5707c235-63be-4f94-8feb-1ae7c83b7926-build-system-configs\") pod \"5707c235-63be-4f94-8feb-1ae7c83b7926\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.597627 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5707c235-63be-4f94-8feb-1ae7c83b7926-build-ca-bundles\") pod \"5707c235-63be-4f94-8feb-1ae7c83b7926\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.597662 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-buildworkdir\") pod \"5707c235-63be-4f94-8feb-1ae7c83b7926\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.597689 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/5707c235-63be-4f94-8feb-1ae7c83b7926-builder-dockercfg-l2p5g-push\") pod \"5707c235-63be-4f94-8feb-1ae7c83b7926\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.597711 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5707c235-63be-4f94-8feb-1ae7c83b7926-buildcachedir\") pod \"5707c235-63be-4f94-8feb-1ae7c83b7926\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.597743 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/5707c235-63be-4f94-8feb-1ae7c83b7926-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"5707c235-63be-4f94-8feb-1ae7c83b7926\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.597779 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-container-storage-root\") pod \"5707c235-63be-4f94-8feb-1ae7c83b7926\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.597805 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5707c235-63be-4f94-8feb-1ae7c83b7926-node-pullsecrets\") pod \"5707c235-63be-4f94-8feb-1ae7c83b7926\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.597831 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-build-blob-cache\") pod \"5707c235-63be-4f94-8feb-1ae7c83b7926\" (UID: \"5707c235-63be-4f94-8feb-1ae7c83b7926\") " Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.598112 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "5707c235-63be-4f94-8feb-1ae7c83b7926" (UID: "5707c235-63be-4f94-8feb-1ae7c83b7926"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.598287 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5707c235-63be-4f94-8feb-1ae7c83b7926-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "5707c235-63be-4f94-8feb-1ae7c83b7926" (UID: "5707c235-63be-4f94-8feb-1ae7c83b7926"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.598305 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5707c235-63be-4f94-8feb-1ae7c83b7926-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "5707c235-63be-4f94-8feb-1ae7c83b7926" (UID: "5707c235-63be-4f94-8feb-1ae7c83b7926"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.598318 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "5707c235-63be-4f94-8feb-1ae7c83b7926" (UID: "5707c235-63be-4f94-8feb-1ae7c83b7926"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.598338 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5707c235-63be-4f94-8feb-1ae7c83b7926-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "5707c235-63be-4f94-8feb-1ae7c83b7926" (UID: "5707c235-63be-4f94-8feb-1ae7c83b7926"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.598630 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "5707c235-63be-4f94-8feb-1ae7c83b7926" (UID: "5707c235-63be-4f94-8feb-1ae7c83b7926"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.598741 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5707c235-63be-4f94-8feb-1ae7c83b7926-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "5707c235-63be-4f94-8feb-1ae7c83b7926" (UID: "5707c235-63be-4f94-8feb-1ae7c83b7926"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.598789 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "5707c235-63be-4f94-8feb-1ae7c83b7926" (UID: "5707c235-63be-4f94-8feb-1ae7c83b7926"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.598874 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5707c235-63be-4f94-8feb-1ae7c83b7926-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "5707c235-63be-4f94-8feb-1ae7c83b7926" (UID: "5707c235-63be-4f94-8feb-1ae7c83b7926"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.603664 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5707c235-63be-4f94-8feb-1ae7c83b7926-kube-api-access-6xl9r" (OuterVolumeSpecName: "kube-api-access-6xl9r") pod "5707c235-63be-4f94-8feb-1ae7c83b7926" (UID: "5707c235-63be-4f94-8feb-1ae7c83b7926"). InnerVolumeSpecName "kube-api-access-6xl9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.604166 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5707c235-63be-4f94-8feb-1ae7c83b7926-builder-dockercfg-l2p5g-pull" (OuterVolumeSpecName: "builder-dockercfg-l2p5g-pull") pod "5707c235-63be-4f94-8feb-1ae7c83b7926" (UID: "5707c235-63be-4f94-8feb-1ae7c83b7926"). InnerVolumeSpecName "builder-dockercfg-l2p5g-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.606776 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5707c235-63be-4f94-8feb-1ae7c83b7926-builder-dockercfg-l2p5g-push" (OuterVolumeSpecName: "builder-dockercfg-l2p5g-push") pod "5707c235-63be-4f94-8feb-1ae7c83b7926" (UID: "5707c235-63be-4f94-8feb-1ae7c83b7926"). InnerVolumeSpecName "builder-dockercfg-l2p5g-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.611809 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5707c235-63be-4f94-8feb-1ae7c83b7926-service-telemetry-framework-index-dockercfg-user-build-volume" (OuterVolumeSpecName: "service-telemetry-framework-index-dockercfg-user-build-volume") pod "5707c235-63be-4f94-8feb-1ae7c83b7926" (UID: "5707c235-63be-4f94-8feb-1ae7c83b7926"). InnerVolumeSpecName "service-telemetry-framework-index-dockercfg-user-build-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.698690 4857 reconciler_common.go:293] "Volume detached for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/5707c235-63be-4f94-8feb-1ae7c83b7926-service-telemetry-framework-index-dockercfg-user-build-volume\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.698727 4857 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.698739 4857 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5707c235-63be-4f94-8feb-1ae7c83b7926-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.698747 4857 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.698756 4857 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5707c235-63be-4f94-8feb-1ae7c83b7926-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.698764 4857 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.698772 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xl9r\" (UniqueName: \"kubernetes.io/projected/5707c235-63be-4f94-8feb-1ae7c83b7926-kube-api-access-6xl9r\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.698781 4857 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/5707c235-63be-4f94-8feb-1ae7c83b7926-builder-dockercfg-l2p5g-pull\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.698790 4857 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5707c235-63be-4f94-8feb-1ae7c83b7926-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.698798 4857 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5707c235-63be-4f94-8feb-1ae7c83b7926-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.698807 4857 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5707c235-63be-4f94-8feb-1ae7c83b7926-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.698816 4857 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5707c235-63be-4f94-8feb-1ae7c83b7926-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:03 crc kubenswrapper[4857]: I0222 00:18:03.698825 4857 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/5707c235-63be-4f94-8feb-1ae7c83b7926-builder-dockercfg-l2p5g-push\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:04 crc kubenswrapper[4857]: I0222 00:18:04.173615 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-framework-index-1-build_5707c235-63be-4f94-8feb-1ae7c83b7926/git-clone/0.log" Feb 22 00:18:04 crc kubenswrapper[4857]: I0222 00:18:04.173914 4857 generic.go:334] "Generic (PLEG): container finished" podID="5707c235-63be-4f94-8feb-1ae7c83b7926" containerID="ac05211744cab7f7ce160169a908907ddb3fb0344b4ec27214992fd339e9f8a5" exitCode=1 Feb 22 00:18:04 crc kubenswrapper[4857]: I0222 00:18:04.173972 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"5707c235-63be-4f94-8feb-1ae7c83b7926","Type":"ContainerDied","Data":"ac05211744cab7f7ce160169a908907ddb3fb0344b4ec27214992fd339e9f8a5"} Feb 22 00:18:04 crc kubenswrapper[4857]: I0222 00:18:04.174022 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-1-build" Feb 22 00:18:04 crc kubenswrapper[4857]: I0222 00:18:04.174054 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-1-build" event={"ID":"5707c235-63be-4f94-8feb-1ae7c83b7926","Type":"ContainerDied","Data":"f1659231b2b5e5f2b1df70da94b8531eceef08d644d709bd886cb494b4356edc"} Feb 22 00:18:04 crc kubenswrapper[4857]: I0222 00:18:04.174068 4857 scope.go:117] "RemoveContainer" containerID="ac05211744cab7f7ce160169a908907ddb3fb0344b4ec27214992fd339e9f8a5" Feb 22 00:18:04 crc kubenswrapper[4857]: I0222 00:18:04.195476 4857 scope.go:117] "RemoveContainer" containerID="ac05211744cab7f7ce160169a908907ddb3fb0344b4ec27214992fd339e9f8a5" Feb 22 00:18:04 crc kubenswrapper[4857]: E0222 00:18:04.195895 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac05211744cab7f7ce160169a908907ddb3fb0344b4ec27214992fd339e9f8a5\": container with ID starting with ac05211744cab7f7ce160169a908907ddb3fb0344b4ec27214992fd339e9f8a5 not found: ID does not exist" containerID="ac05211744cab7f7ce160169a908907ddb3fb0344b4ec27214992fd339e9f8a5" Feb 22 00:18:04 crc kubenswrapper[4857]: I0222 00:18:04.195942 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac05211744cab7f7ce160169a908907ddb3fb0344b4ec27214992fd339e9f8a5"} err="failed to get container status \"ac05211744cab7f7ce160169a908907ddb3fb0344b4ec27214992fd339e9f8a5\": rpc error: code = NotFound desc = could not find container \"ac05211744cab7f7ce160169a908907ddb3fb0344b4ec27214992fd339e9f8a5\": container with ID starting with ac05211744cab7f7ce160169a908907ddb3fb0344b4ec27214992fd339e9f8a5 not found: ID does not exist" Feb 22 00:18:04 crc kubenswrapper[4857]: I0222 00:18:04.212813 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-framework-index-1-build"] Feb 22 00:18:04 crc kubenswrapper[4857]: I0222 00:18:04.221642 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-framework-index-1-build"] Feb 22 00:18:04 crc kubenswrapper[4857]: I0222 00:18:04.808160 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-545d4d4674-6nr72"] Feb 22 00:18:04 crc kubenswrapper[4857]: E0222 00:18:04.808789 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5707c235-63be-4f94-8feb-1ae7c83b7926" containerName="git-clone" Feb 22 00:18:04 crc kubenswrapper[4857]: I0222 00:18:04.808817 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="5707c235-63be-4f94-8feb-1ae7c83b7926" containerName="git-clone" Feb 22 00:18:04 crc kubenswrapper[4857]: I0222 00:18:04.809142 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="5707c235-63be-4f94-8feb-1ae7c83b7926" containerName="git-clone" Feb 22 00:18:04 crc kubenswrapper[4857]: I0222 00:18:04.810318 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-6nr72" Feb 22 00:18:04 crc kubenswrapper[4857]: I0222 00:18:04.816267 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-6nr72"] Feb 22 00:18:04 crc kubenswrapper[4857]: I0222 00:18:04.816746 4857 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-v7qcp" Feb 22 00:18:04 crc kubenswrapper[4857]: I0222 00:18:04.928285 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/63f9c6da-3d8d-4ae7-b619-52db9af8a21f-bound-sa-token\") pod \"cert-manager-545d4d4674-6nr72\" (UID: \"63f9c6da-3d8d-4ae7-b619-52db9af8a21f\") " pod="cert-manager/cert-manager-545d4d4674-6nr72" Feb 22 00:18:04 crc kubenswrapper[4857]: I0222 00:18:04.928358 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4bfg\" (UniqueName: \"kubernetes.io/projected/63f9c6da-3d8d-4ae7-b619-52db9af8a21f-kube-api-access-f4bfg\") pod \"cert-manager-545d4d4674-6nr72\" (UID: \"63f9c6da-3d8d-4ae7-b619-52db9af8a21f\") " pod="cert-manager/cert-manager-545d4d4674-6nr72" Feb 22 00:18:05 crc kubenswrapper[4857]: I0222 00:18:05.029672 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4bfg\" (UniqueName: \"kubernetes.io/projected/63f9c6da-3d8d-4ae7-b619-52db9af8a21f-kube-api-access-f4bfg\") pod \"cert-manager-545d4d4674-6nr72\" (UID: \"63f9c6da-3d8d-4ae7-b619-52db9af8a21f\") " pod="cert-manager/cert-manager-545d4d4674-6nr72" Feb 22 00:18:05 crc kubenswrapper[4857]: I0222 00:18:05.030074 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/63f9c6da-3d8d-4ae7-b619-52db9af8a21f-bound-sa-token\") pod \"cert-manager-545d4d4674-6nr72\" (UID: \"63f9c6da-3d8d-4ae7-b619-52db9af8a21f\") " pod="cert-manager/cert-manager-545d4d4674-6nr72" Feb 22 00:18:05 crc kubenswrapper[4857]: I0222 00:18:05.052966 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/63f9c6da-3d8d-4ae7-b619-52db9af8a21f-bound-sa-token\") pod \"cert-manager-545d4d4674-6nr72\" (UID: \"63f9c6da-3d8d-4ae7-b619-52db9af8a21f\") " pod="cert-manager/cert-manager-545d4d4674-6nr72" Feb 22 00:18:05 crc kubenswrapper[4857]: I0222 00:18:05.053370 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4bfg\" (UniqueName: \"kubernetes.io/projected/63f9c6da-3d8d-4ae7-b619-52db9af8a21f-kube-api-access-f4bfg\") pod \"cert-manager-545d4d4674-6nr72\" (UID: \"63f9c6da-3d8d-4ae7-b619-52db9af8a21f\") " pod="cert-manager/cert-manager-545d4d4674-6nr72" Feb 22 00:18:05 crc kubenswrapper[4857]: I0222 00:18:05.084809 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5707c235-63be-4f94-8feb-1ae7c83b7926" path="/var/lib/kubelet/pods/5707c235-63be-4f94-8feb-1ae7c83b7926/volumes" Feb 22 00:18:05 crc kubenswrapper[4857]: I0222 00:18:05.153798 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-6nr72" Feb 22 00:18:05 crc kubenswrapper[4857]: I0222 00:18:05.380518 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-6nr72"] Feb 22 00:18:05 crc kubenswrapper[4857]: I0222 00:18:05.763811 4857 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="ee783ac5-b162-4821-87ff-09dbc60e98f1" containerName="elasticsearch" probeResult="failure" output=< Feb 22 00:18:05 crc kubenswrapper[4857]: {"timestamp": "2026-02-22T00:18:05+00:00", "message": "readiness probe failed", "curl_rc": "7"} Feb 22 00:18:05 crc kubenswrapper[4857]: > Feb 22 00:18:06 crc kubenswrapper[4857]: I0222 00:18:06.199906 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-6nr72" event={"ID":"63f9c6da-3d8d-4ae7-b619-52db9af8a21f","Type":"ContainerStarted","Data":"07b8c2d04212e6318176f87bc49f6e28dff525b78e9109306c361fe43da992ad"} Feb 22 00:18:06 crc kubenswrapper[4857]: I0222 00:18:06.200277 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-6nr72" event={"ID":"63f9c6da-3d8d-4ae7-b619-52db9af8a21f","Type":"ContainerStarted","Data":"c2c5f348269f98a89bf05c1428209811b43af81eec229cceb8a99f41ac6ca4d2"} Feb 22 00:18:06 crc kubenswrapper[4857]: I0222 00:18:06.214666 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-545d4d4674-6nr72" podStartSLOduration=2.214645178 podStartE2EDuration="2.214645178s" podCreationTimestamp="2026-02-22 00:18:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:18:06.212402449 +0000 UTC m=+693.851131712" watchObservedRunningTime="2026-02-22 00:18:06.214645178 +0000 UTC m=+693.853374431" Feb 22 00:18:10 crc kubenswrapper[4857]: I0222 00:18:10.428739 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:18:10 crc kubenswrapper[4857]: I0222 00:18:10.429070 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:18:11 crc kubenswrapper[4857]: I0222 00:18:11.221557 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.825038 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-framework-index-2-build"] Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.826463 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.828639 4857 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-l2p5g" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.829205 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-2-sys-config" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.829243 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-2-ca" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.829659 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-2-global-ca" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.832361 4857 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"service-telemetry-framework-index-dockercfg" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.841248 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/a00214d3-a2f1-4398-afd8-b68e1b2707f4-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.842622 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/a00214d3-a2f1-4398-afd8-b68e1b2707f4-builder-dockercfg-l2p5g-pull\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.842749 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-blob-cache\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.842894 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.843019 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a00214d3-a2f1-4398-afd8-b68e1b2707f4-node-pullsecrets\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.843185 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a00214d3-a2f1-4398-afd8-b68e1b2707f4-buildcachedir\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.843300 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-855wm\" (UniqueName: \"kubernetes.io/projected/a00214d3-a2f1-4398-afd8-b68e1b2707f4-kube-api-access-855wm\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.843400 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-system-configs\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.843526 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-container-storage-root\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.843658 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-buildworkdir\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.843755 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/a00214d3-a2f1-4398-afd8-b68e1b2707f4-builder-dockercfg-l2p5g-push\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.843859 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-container-storage-run\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.843961 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-ca-bundles\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.845724 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-framework-index-2-build"] Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.945941 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-buildworkdir\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.945998 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/a00214d3-a2f1-4398-afd8-b68e1b2707f4-builder-dockercfg-l2p5g-push\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.946040 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-container-storage-run\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.946096 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-ca-bundles\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.946131 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/a00214d3-a2f1-4398-afd8-b68e1b2707f4-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.946161 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/a00214d3-a2f1-4398-afd8-b68e1b2707f4-builder-dockercfg-l2p5g-pull\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.946183 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-blob-cache\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.946205 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.946230 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a00214d3-a2f1-4398-afd8-b68e1b2707f4-node-pullsecrets\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.946261 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a00214d3-a2f1-4398-afd8-b68e1b2707f4-buildcachedir\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.946298 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-855wm\" (UniqueName: \"kubernetes.io/projected/a00214d3-a2f1-4398-afd8-b68e1b2707f4-kube-api-access-855wm\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.946326 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-system-configs\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.946364 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-container-storage-root\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.947168 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a00214d3-a2f1-4398-afd8-b68e1b2707f4-node-pullsecrets\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.947200 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-blob-cache\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.947291 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a00214d3-a2f1-4398-afd8-b68e1b2707f4-buildcachedir\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.947493 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-container-storage-root\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.947554 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-ca-bundles\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.947667 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.947709 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-buildworkdir\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.947796 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-container-storage-run\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.948170 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-system-configs\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.958668 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/a00214d3-a2f1-4398-afd8-b68e1b2707f4-builder-dockercfg-l2p5g-push\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.958840 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/a00214d3-a2f1-4398-afd8-b68e1b2707f4-builder-dockercfg-l2p5g-pull\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.958688 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/a00214d3-a2f1-4398-afd8-b68e1b2707f4-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:13 crc kubenswrapper[4857]: I0222 00:18:13.965743 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-855wm\" (UniqueName: \"kubernetes.io/projected/a00214d3-a2f1-4398-afd8-b68e1b2707f4-kube-api-access-855wm\") pod \"service-telemetry-framework-index-2-build\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:14 crc kubenswrapper[4857]: I0222 00:18:14.148358 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:14 crc kubenswrapper[4857]: I0222 00:18:14.350740 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-framework-index-2-build"] Feb 22 00:18:15 crc kubenswrapper[4857]: I0222 00:18:15.269903 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-2-build" event={"ID":"a00214d3-a2f1-4398-afd8-b68e1b2707f4","Type":"ContainerStarted","Data":"c05ed3a56e38fd572df261ae432775ba7e439590a001a728bfe01553d0a2313f"} Feb 22 00:18:15 crc kubenswrapper[4857]: I0222 00:18:15.270321 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-2-build" event={"ID":"a00214d3-a2f1-4398-afd8-b68e1b2707f4","Type":"ContainerStarted","Data":"0245df303b52c64d3aeeb916f49061c147ef51be4d424256f11b651e38e08af7"} Feb 22 00:18:15 crc kubenswrapper[4857]: E0222 00:18:15.329669 4857 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=8528648506435008715, SKID=, AKID=49:B4:28:90:5C:47:E0:09:27:27:EB:F1:16:CB:58:04:C2:1D:C0:A3 failed: x509: certificate signed by unknown authority" Feb 22 00:18:16 crc kubenswrapper[4857]: I0222 00:18:16.355531 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-framework-index-2-build"] Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.285740 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-framework-index-2-build" podUID="a00214d3-a2f1-4398-afd8-b68e1b2707f4" containerName="git-clone" containerID="cri-o://c05ed3a56e38fd572df261ae432775ba7e439590a001a728bfe01553d0a2313f" gracePeriod=30 Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.689341 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-framework-index-2-build_a00214d3-a2f1-4398-afd8-b68e1b2707f4/git-clone/0.log" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.689424 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.795581 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a00214d3-a2f1-4398-afd8-b68e1b2707f4-buildcachedir\") pod \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.795674 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-container-storage-root\") pod \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.795710 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-system-configs\") pod \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.795742 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a00214d3-a2f1-4398-afd8-b68e1b2707f4-node-pullsecrets\") pod \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.795762 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-ca-bundles\") pod \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.795790 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/a00214d3-a2f1-4398-afd8-b68e1b2707f4-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.795834 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-855wm\" (UniqueName: \"kubernetes.io/projected/a00214d3-a2f1-4398-afd8-b68e1b2707f4-kube-api-access-855wm\") pod \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.795829 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a00214d3-a2f1-4398-afd8-b68e1b2707f4-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "a00214d3-a2f1-4398-afd8-b68e1b2707f4" (UID: "a00214d3-a2f1-4398-afd8-b68e1b2707f4"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.795855 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/a00214d3-a2f1-4398-afd8-b68e1b2707f4-builder-dockercfg-l2p5g-pull\") pod \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.795942 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-proxy-ca-bundles\") pod \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.795985 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-blob-cache\") pod \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.796088 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-container-storage-run\") pod \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.796115 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-buildworkdir\") pod \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.796116 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "a00214d3-a2f1-4398-afd8-b68e1b2707f4" (UID: "a00214d3-a2f1-4398-afd8-b68e1b2707f4"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.796138 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/a00214d3-a2f1-4398-afd8-b68e1b2707f4-builder-dockercfg-l2p5g-push\") pod \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\" (UID: \"a00214d3-a2f1-4398-afd8-b68e1b2707f4\") " Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.796192 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "a00214d3-a2f1-4398-afd8-b68e1b2707f4" (UID: "a00214d3-a2f1-4398-afd8-b68e1b2707f4"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.796347 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "a00214d3-a2f1-4398-afd8-b68e1b2707f4" (UID: "a00214d3-a2f1-4398-afd8-b68e1b2707f4"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.796612 4857 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.796634 4857 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.796649 4857 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.796663 4857 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a00214d3-a2f1-4398-afd8-b68e1b2707f4-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.796714 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "a00214d3-a2f1-4398-afd8-b68e1b2707f4" (UID: "a00214d3-a2f1-4398-afd8-b68e1b2707f4"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.797095 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "a00214d3-a2f1-4398-afd8-b68e1b2707f4" (UID: "a00214d3-a2f1-4398-afd8-b68e1b2707f4"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.797148 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "a00214d3-a2f1-4398-afd8-b68e1b2707f4" (UID: "a00214d3-a2f1-4398-afd8-b68e1b2707f4"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.797495 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a00214d3-a2f1-4398-afd8-b68e1b2707f4-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "a00214d3-a2f1-4398-afd8-b68e1b2707f4" (UID: "a00214d3-a2f1-4398-afd8-b68e1b2707f4"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.797604 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "a00214d3-a2f1-4398-afd8-b68e1b2707f4" (UID: "a00214d3-a2f1-4398-afd8-b68e1b2707f4"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.801292 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a00214d3-a2f1-4398-afd8-b68e1b2707f4-service-telemetry-framework-index-dockercfg-user-build-volume" (OuterVolumeSpecName: "service-telemetry-framework-index-dockercfg-user-build-volume") pod "a00214d3-a2f1-4398-afd8-b68e1b2707f4" (UID: "a00214d3-a2f1-4398-afd8-b68e1b2707f4"). InnerVolumeSpecName "service-telemetry-framework-index-dockercfg-user-build-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.801378 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a00214d3-a2f1-4398-afd8-b68e1b2707f4-kube-api-access-855wm" (OuterVolumeSpecName: "kube-api-access-855wm") pod "a00214d3-a2f1-4398-afd8-b68e1b2707f4" (UID: "a00214d3-a2f1-4398-afd8-b68e1b2707f4"). InnerVolumeSpecName "kube-api-access-855wm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.801576 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a00214d3-a2f1-4398-afd8-b68e1b2707f4-builder-dockercfg-l2p5g-push" (OuterVolumeSpecName: "builder-dockercfg-l2p5g-push") pod "a00214d3-a2f1-4398-afd8-b68e1b2707f4" (UID: "a00214d3-a2f1-4398-afd8-b68e1b2707f4"). InnerVolumeSpecName "builder-dockercfg-l2p5g-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.805445 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a00214d3-a2f1-4398-afd8-b68e1b2707f4-builder-dockercfg-l2p5g-pull" (OuterVolumeSpecName: "builder-dockercfg-l2p5g-pull") pod "a00214d3-a2f1-4398-afd8-b68e1b2707f4" (UID: "a00214d3-a2f1-4398-afd8-b68e1b2707f4"). InnerVolumeSpecName "builder-dockercfg-l2p5g-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.898157 4857 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a00214d3-a2f1-4398-afd8-b68e1b2707f4-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.898460 4857 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.898576 4857 reconciler_common.go:293] "Volume detached for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/a00214d3-a2f1-4398-afd8-b68e1b2707f4-service-telemetry-framework-index-dockercfg-user-build-volume\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.898742 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-855wm\" (UniqueName: \"kubernetes.io/projected/a00214d3-a2f1-4398-afd8-b68e1b2707f4-kube-api-access-855wm\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.898859 4857 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/a00214d3-a2f1-4398-afd8-b68e1b2707f4-builder-dockercfg-l2p5g-pull\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.898972 4857 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a00214d3-a2f1-4398-afd8-b68e1b2707f4-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.899089 4857 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.899186 4857 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a00214d3-a2f1-4398-afd8-b68e1b2707f4-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:17 crc kubenswrapper[4857]: I0222 00:18:17.899280 4857 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/a00214d3-a2f1-4398-afd8-b68e1b2707f4-builder-dockercfg-l2p5g-push\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:18 crc kubenswrapper[4857]: I0222 00:18:18.294888 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-framework-index-2-build_a00214d3-a2f1-4398-afd8-b68e1b2707f4/git-clone/0.log" Feb 22 00:18:18 crc kubenswrapper[4857]: I0222 00:18:18.294943 4857 generic.go:334] "Generic (PLEG): container finished" podID="a00214d3-a2f1-4398-afd8-b68e1b2707f4" containerID="c05ed3a56e38fd572df261ae432775ba7e439590a001a728bfe01553d0a2313f" exitCode=1 Feb 22 00:18:18 crc kubenswrapper[4857]: I0222 00:18:18.294974 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-2-build" event={"ID":"a00214d3-a2f1-4398-afd8-b68e1b2707f4","Type":"ContainerDied","Data":"c05ed3a56e38fd572df261ae432775ba7e439590a001a728bfe01553d0a2313f"} Feb 22 00:18:18 crc kubenswrapper[4857]: I0222 00:18:18.295005 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-2-build" event={"ID":"a00214d3-a2f1-4398-afd8-b68e1b2707f4","Type":"ContainerDied","Data":"0245df303b52c64d3aeeb916f49061c147ef51be4d424256f11b651e38e08af7"} Feb 22 00:18:18 crc kubenswrapper[4857]: I0222 00:18:18.295023 4857 scope.go:117] "RemoveContainer" containerID="c05ed3a56e38fd572df261ae432775ba7e439590a001a728bfe01553d0a2313f" Feb 22 00:18:18 crc kubenswrapper[4857]: I0222 00:18:18.295351 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-2-build" Feb 22 00:18:18 crc kubenswrapper[4857]: I0222 00:18:18.316642 4857 scope.go:117] "RemoveContainer" containerID="c05ed3a56e38fd572df261ae432775ba7e439590a001a728bfe01553d0a2313f" Feb 22 00:18:18 crc kubenswrapper[4857]: E0222 00:18:18.317064 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c05ed3a56e38fd572df261ae432775ba7e439590a001a728bfe01553d0a2313f\": container with ID starting with c05ed3a56e38fd572df261ae432775ba7e439590a001a728bfe01553d0a2313f not found: ID does not exist" containerID="c05ed3a56e38fd572df261ae432775ba7e439590a001a728bfe01553d0a2313f" Feb 22 00:18:18 crc kubenswrapper[4857]: I0222 00:18:18.317117 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c05ed3a56e38fd572df261ae432775ba7e439590a001a728bfe01553d0a2313f"} err="failed to get container status \"c05ed3a56e38fd572df261ae432775ba7e439590a001a728bfe01553d0a2313f\": rpc error: code = NotFound desc = could not find container \"c05ed3a56e38fd572df261ae432775ba7e439590a001a728bfe01553d0a2313f\": container with ID starting with c05ed3a56e38fd572df261ae432775ba7e439590a001a728bfe01553d0a2313f not found: ID does not exist" Feb 22 00:18:18 crc kubenswrapper[4857]: I0222 00:18:18.336609 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-framework-index-2-build"] Feb 22 00:18:18 crc kubenswrapper[4857]: I0222 00:18:18.345325 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-framework-index-2-build"] Feb 22 00:18:19 crc kubenswrapper[4857]: I0222 00:18:19.103629 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a00214d3-a2f1-4398-afd8-b68e1b2707f4" path="/var/lib/kubelet/pods/a00214d3-a2f1-4398-afd8-b68e1b2707f4/volumes" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.774587 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-framework-index-3-build"] Feb 22 00:18:27 crc kubenswrapper[4857]: E0222 00:18:27.775777 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a00214d3-a2f1-4398-afd8-b68e1b2707f4" containerName="git-clone" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.775796 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="a00214d3-a2f1-4398-afd8-b68e1b2707f4" containerName="git-clone" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.775947 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="a00214d3-a2f1-4398-afd8-b68e1b2707f4" containerName="git-clone" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.776890 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.779054 4857 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"service-telemetry-framework-index-dockercfg" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.779099 4857 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-l2p5g" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.779431 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-3-ca" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.779944 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-3-sys-config" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.782853 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-3-global-ca" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.791737 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-framework-index-3-build"] Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.939548 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-container-storage-run\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.939633 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-container-storage-root\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.939684 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/66b9e066-0a58-41f1-bcc2-6faee3756853-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.939727 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-build-blob-cache\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.939779 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/66b9e066-0a58-41f1-bcc2-6faee3756853-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.939821 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/66b9e066-0a58-41f1-bcc2-6faee3756853-buildcachedir\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.939868 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/66b9e066-0a58-41f1-bcc2-6faee3756853-builder-dockercfg-l2p5g-pull\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.939917 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7d9l\" (UniqueName: \"kubernetes.io/projected/66b9e066-0a58-41f1-bcc2-6faee3756853-kube-api-access-z7d9l\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.939944 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-buildworkdir\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.939990 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/66b9e066-0a58-41f1-bcc2-6faee3756853-builder-dockercfg-l2p5g-push\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.940093 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/66b9e066-0a58-41f1-bcc2-6faee3756853-build-ca-bundles\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.940131 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/66b9e066-0a58-41f1-bcc2-6faee3756853-node-pullsecrets\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:27 crc kubenswrapper[4857]: I0222 00:18:27.940169 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/66b9e066-0a58-41f1-bcc2-6faee3756853-build-system-configs\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.042334 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-container-storage-run\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.042434 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-container-storage-root\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.042495 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-build-blob-cache\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.042545 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/66b9e066-0a58-41f1-bcc2-6faee3756853-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.042615 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/66b9e066-0a58-41f1-bcc2-6faee3756853-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.042676 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/66b9e066-0a58-41f1-bcc2-6faee3756853-buildcachedir\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.042747 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/66b9e066-0a58-41f1-bcc2-6faee3756853-builder-dockercfg-l2p5g-pull\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.042823 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-buildworkdir\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.042913 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/66b9e066-0a58-41f1-bcc2-6faee3756853-buildcachedir\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.042951 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7d9l\" (UniqueName: \"kubernetes.io/projected/66b9e066-0a58-41f1-bcc2-6faee3756853-kube-api-access-z7d9l\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.042828 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-container-storage-run\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.043013 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-container-storage-root\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.043077 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/66b9e066-0a58-41f1-bcc2-6faee3756853-builder-dockercfg-l2p5g-push\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.043267 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/66b9e066-0a58-41f1-bcc2-6faee3756853-build-ca-bundles\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.043330 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/66b9e066-0a58-41f1-bcc2-6faee3756853-node-pullsecrets\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.043395 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/66b9e066-0a58-41f1-bcc2-6faee3756853-build-system-configs\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.043458 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-buildworkdir\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.043488 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/66b9e066-0a58-41f1-bcc2-6faee3756853-node-pullsecrets\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.043718 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/66b9e066-0a58-41f1-bcc2-6faee3756853-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.043778 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-build-blob-cache\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.044451 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/66b9e066-0a58-41f1-bcc2-6faee3756853-build-system-configs\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.045690 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/66b9e066-0a58-41f1-bcc2-6faee3756853-build-ca-bundles\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.051706 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/66b9e066-0a58-41f1-bcc2-6faee3756853-builder-dockercfg-l2p5g-push\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.051809 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/66b9e066-0a58-41f1-bcc2-6faee3756853-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.054629 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/66b9e066-0a58-41f1-bcc2-6faee3756853-builder-dockercfg-l2p5g-pull\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.083831 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7d9l\" (UniqueName: \"kubernetes.io/projected/66b9e066-0a58-41f1-bcc2-6faee3756853-kube-api-access-z7d9l\") pod \"service-telemetry-framework-index-3-build\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.092658 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:28 crc kubenswrapper[4857]: I0222 00:18:28.531869 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-framework-index-3-build"] Feb 22 00:18:29 crc kubenswrapper[4857]: I0222 00:18:29.373800 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-3-build" event={"ID":"66b9e066-0a58-41f1-bcc2-6faee3756853","Type":"ContainerStarted","Data":"5c36c459ed871c98e8672b51a073c96a77a1677828f38420a8a957bb72ec5865"} Feb 22 00:18:29 crc kubenswrapper[4857]: I0222 00:18:29.374137 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-3-build" event={"ID":"66b9e066-0a58-41f1-bcc2-6faee3756853","Type":"ContainerStarted","Data":"0712d3ecd7ceabff05e695d646745e5d7d4efa9ec97452565fa2d1a8e3974f8d"} Feb 22 00:18:29 crc kubenswrapper[4857]: E0222 00:18:29.437116 4857 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=8528648506435008715, SKID=, AKID=49:B4:28:90:5C:47:E0:09:27:27:EB:F1:16:CB:58:04:C2:1D:C0:A3 failed: x509: certificate signed by unknown authority" Feb 22 00:18:30 crc kubenswrapper[4857]: I0222 00:18:30.465685 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-framework-index-3-build"] Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.386564 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-framework-index-3-build" podUID="66b9e066-0a58-41f1-bcc2-6faee3756853" containerName="git-clone" containerID="cri-o://5c36c459ed871c98e8672b51a073c96a77a1677828f38420a8a957bb72ec5865" gracePeriod=30 Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.755103 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-framework-index-3-build_66b9e066-0a58-41f1-bcc2-6faee3756853/git-clone/0.log" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.755432 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.890903 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/66b9e066-0a58-41f1-bcc2-6faee3756853-build-ca-bundles\") pod \"66b9e066-0a58-41f1-bcc2-6faee3756853\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.890995 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/66b9e066-0a58-41f1-bcc2-6faee3756853-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"66b9e066-0a58-41f1-bcc2-6faee3756853\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.891017 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/66b9e066-0a58-41f1-bcc2-6faee3756853-build-proxy-ca-bundles\") pod \"66b9e066-0a58-41f1-bcc2-6faee3756853\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.891094 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/66b9e066-0a58-41f1-bcc2-6faee3756853-builder-dockercfg-l2p5g-pull\") pod \"66b9e066-0a58-41f1-bcc2-6faee3756853\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.891115 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-container-storage-root\") pod \"66b9e066-0a58-41f1-bcc2-6faee3756853\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.891133 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-buildworkdir\") pod \"66b9e066-0a58-41f1-bcc2-6faee3756853\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.891175 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7d9l\" (UniqueName: \"kubernetes.io/projected/66b9e066-0a58-41f1-bcc2-6faee3756853-kube-api-access-z7d9l\") pod \"66b9e066-0a58-41f1-bcc2-6faee3756853\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.891256 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-build-blob-cache\") pod \"66b9e066-0a58-41f1-bcc2-6faee3756853\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.891288 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/66b9e066-0a58-41f1-bcc2-6faee3756853-node-pullsecrets\") pod \"66b9e066-0a58-41f1-bcc2-6faee3756853\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.891309 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/66b9e066-0a58-41f1-bcc2-6faee3756853-build-system-configs\") pod \"66b9e066-0a58-41f1-bcc2-6faee3756853\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.891336 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/66b9e066-0a58-41f1-bcc2-6faee3756853-buildcachedir\") pod \"66b9e066-0a58-41f1-bcc2-6faee3756853\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.891354 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/66b9e066-0a58-41f1-bcc2-6faee3756853-builder-dockercfg-l2p5g-push\") pod \"66b9e066-0a58-41f1-bcc2-6faee3756853\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.891374 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-container-storage-run\") pod \"66b9e066-0a58-41f1-bcc2-6faee3756853\" (UID: \"66b9e066-0a58-41f1-bcc2-6faee3756853\") " Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.891765 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66b9e066-0a58-41f1-bcc2-6faee3756853-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "66b9e066-0a58-41f1-bcc2-6faee3756853" (UID: "66b9e066-0a58-41f1-bcc2-6faee3756853"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.891775 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66b9e066-0a58-41f1-bcc2-6faee3756853-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "66b9e066-0a58-41f1-bcc2-6faee3756853" (UID: "66b9e066-0a58-41f1-bcc2-6faee3756853"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.892364 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66b9e066-0a58-41f1-bcc2-6faee3756853-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "66b9e066-0a58-41f1-bcc2-6faee3756853" (UID: "66b9e066-0a58-41f1-bcc2-6faee3756853"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.892388 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "66b9e066-0a58-41f1-bcc2-6faee3756853" (UID: "66b9e066-0a58-41f1-bcc2-6faee3756853"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.892494 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "66b9e066-0a58-41f1-bcc2-6faee3756853" (UID: "66b9e066-0a58-41f1-bcc2-6faee3756853"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.892441 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "66b9e066-0a58-41f1-bcc2-6faee3756853" (UID: "66b9e066-0a58-41f1-bcc2-6faee3756853"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.892525 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "66b9e066-0a58-41f1-bcc2-6faee3756853" (UID: "66b9e066-0a58-41f1-bcc2-6faee3756853"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.892543 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66b9e066-0a58-41f1-bcc2-6faee3756853-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "66b9e066-0a58-41f1-bcc2-6faee3756853" (UID: "66b9e066-0a58-41f1-bcc2-6faee3756853"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.892731 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66b9e066-0a58-41f1-bcc2-6faee3756853-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "66b9e066-0a58-41f1-bcc2-6faee3756853" (UID: "66b9e066-0a58-41f1-bcc2-6faee3756853"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.897080 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66b9e066-0a58-41f1-bcc2-6faee3756853-kube-api-access-z7d9l" (OuterVolumeSpecName: "kube-api-access-z7d9l") pod "66b9e066-0a58-41f1-bcc2-6faee3756853" (UID: "66b9e066-0a58-41f1-bcc2-6faee3756853"). InnerVolumeSpecName "kube-api-access-z7d9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.898167 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66b9e066-0a58-41f1-bcc2-6faee3756853-builder-dockercfg-l2p5g-push" (OuterVolumeSpecName: "builder-dockercfg-l2p5g-push") pod "66b9e066-0a58-41f1-bcc2-6faee3756853" (UID: "66b9e066-0a58-41f1-bcc2-6faee3756853"). InnerVolumeSpecName "builder-dockercfg-l2p5g-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.898199 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66b9e066-0a58-41f1-bcc2-6faee3756853-builder-dockercfg-l2p5g-pull" (OuterVolumeSpecName: "builder-dockercfg-l2p5g-pull") pod "66b9e066-0a58-41f1-bcc2-6faee3756853" (UID: "66b9e066-0a58-41f1-bcc2-6faee3756853"). InnerVolumeSpecName "builder-dockercfg-l2p5g-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.903193 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66b9e066-0a58-41f1-bcc2-6faee3756853-service-telemetry-framework-index-dockercfg-user-build-volume" (OuterVolumeSpecName: "service-telemetry-framework-index-dockercfg-user-build-volume") pod "66b9e066-0a58-41f1-bcc2-6faee3756853" (UID: "66b9e066-0a58-41f1-bcc2-6faee3756853"). InnerVolumeSpecName "service-telemetry-framework-index-dockercfg-user-build-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.992992 4857 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.993097 4857 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/66b9e066-0a58-41f1-bcc2-6faee3756853-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.993127 4857 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/66b9e066-0a58-41f1-bcc2-6faee3756853-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.993154 4857 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/66b9e066-0a58-41f1-bcc2-6faee3756853-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.993175 4857 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/66b9e066-0a58-41f1-bcc2-6faee3756853-builder-dockercfg-l2p5g-push\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.993200 4857 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.993223 4857 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/66b9e066-0a58-41f1-bcc2-6faee3756853-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.993249 4857 reconciler_common.go:293] "Volume detached for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/66b9e066-0a58-41f1-bcc2-6faee3756853-service-telemetry-framework-index-dockercfg-user-build-volume\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.993277 4857 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/66b9e066-0a58-41f1-bcc2-6faee3756853-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.993300 4857 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/66b9e066-0a58-41f1-bcc2-6faee3756853-builder-dockercfg-l2p5g-pull\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.993319 4857 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.993342 4857 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/66b9e066-0a58-41f1-bcc2-6faee3756853-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:31 crc kubenswrapper[4857]: I0222 00:18:31.993368 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7d9l\" (UniqueName: \"kubernetes.io/projected/66b9e066-0a58-41f1-bcc2-6faee3756853-kube-api-access-z7d9l\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:32 crc kubenswrapper[4857]: I0222 00:18:32.395176 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-framework-index-3-build_66b9e066-0a58-41f1-bcc2-6faee3756853/git-clone/0.log" Feb 22 00:18:32 crc kubenswrapper[4857]: I0222 00:18:32.395227 4857 generic.go:334] "Generic (PLEG): container finished" podID="66b9e066-0a58-41f1-bcc2-6faee3756853" containerID="5c36c459ed871c98e8672b51a073c96a77a1677828f38420a8a957bb72ec5865" exitCode=1 Feb 22 00:18:32 crc kubenswrapper[4857]: I0222 00:18:32.395258 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-3-build" event={"ID":"66b9e066-0a58-41f1-bcc2-6faee3756853","Type":"ContainerDied","Data":"5c36c459ed871c98e8672b51a073c96a77a1677828f38420a8a957bb72ec5865"} Feb 22 00:18:32 crc kubenswrapper[4857]: I0222 00:18:32.395289 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-3-build" event={"ID":"66b9e066-0a58-41f1-bcc2-6faee3756853","Type":"ContainerDied","Data":"0712d3ecd7ceabff05e695d646745e5d7d4efa9ec97452565fa2d1a8e3974f8d"} Feb 22 00:18:32 crc kubenswrapper[4857]: I0222 00:18:32.395308 4857 scope.go:117] "RemoveContainer" containerID="5c36c459ed871c98e8672b51a073c96a77a1677828f38420a8a957bb72ec5865" Feb 22 00:18:32 crc kubenswrapper[4857]: I0222 00:18:32.395332 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-3-build" Feb 22 00:18:32 crc kubenswrapper[4857]: I0222 00:18:32.410949 4857 scope.go:117] "RemoveContainer" containerID="5c36c459ed871c98e8672b51a073c96a77a1677828f38420a8a957bb72ec5865" Feb 22 00:18:32 crc kubenswrapper[4857]: E0222 00:18:32.411468 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c36c459ed871c98e8672b51a073c96a77a1677828f38420a8a957bb72ec5865\": container with ID starting with 5c36c459ed871c98e8672b51a073c96a77a1677828f38420a8a957bb72ec5865 not found: ID does not exist" containerID="5c36c459ed871c98e8672b51a073c96a77a1677828f38420a8a957bb72ec5865" Feb 22 00:18:32 crc kubenswrapper[4857]: I0222 00:18:32.411538 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c36c459ed871c98e8672b51a073c96a77a1677828f38420a8a957bb72ec5865"} err="failed to get container status \"5c36c459ed871c98e8672b51a073c96a77a1677828f38420a8a957bb72ec5865\": rpc error: code = NotFound desc = could not find container \"5c36c459ed871c98e8672b51a073c96a77a1677828f38420a8a957bb72ec5865\": container with ID starting with 5c36c459ed871c98e8672b51a073c96a77a1677828f38420a8a957bb72ec5865 not found: ID does not exist" Feb 22 00:18:32 crc kubenswrapper[4857]: I0222 00:18:32.444429 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-framework-index-3-build"] Feb 22 00:18:32 crc kubenswrapper[4857]: I0222 00:18:32.450077 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-framework-index-3-build"] Feb 22 00:18:33 crc kubenswrapper[4857]: I0222 00:18:33.085654 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66b9e066-0a58-41f1-bcc2-6faee3756853" path="/var/lib/kubelet/pods/66b9e066-0a58-41f1-bcc2-6faee3756853/volumes" Feb 22 00:18:40 crc kubenswrapper[4857]: I0222 00:18:40.428224 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:18:40 crc kubenswrapper[4857]: I0222 00:18:40.428982 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.887095 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-framework-index-4-build"] Feb 22 00:18:41 crc kubenswrapper[4857]: E0222 00:18:41.887662 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66b9e066-0a58-41f1-bcc2-6faee3756853" containerName="git-clone" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.887678 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="66b9e066-0a58-41f1-bcc2-6faee3756853" containerName="git-clone" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.887807 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="66b9e066-0a58-41f1-bcc2-6faee3756853" containerName="git-clone" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.888694 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.890227 4857 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"service-telemetry-framework-index-dockercfg" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.890989 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-4-sys-config" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.891177 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-4-ca" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.891486 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-framework-index-4-global-ca" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.891627 4857 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-l2p5g" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.924972 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-framework-index-4-build"] Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.946834 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-ca-bundles\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.946898 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-buildworkdir\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.946923 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-container-storage-run\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.946958 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-system-configs\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.946980 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-blob-cache\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.947006 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-builder-dockercfg-l2p5g-push\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.947023 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.947144 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-builder-dockercfg-l2p5g-pull\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.947184 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-container-storage-root\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.947207 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-node-pullsecrets\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.947234 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-buildcachedir\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.947262 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:41 crc kubenswrapper[4857]: I0222 00:18:41.947284 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf6b5\" (UniqueName: \"kubernetes.io/projected/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-kube-api-access-tf6b5\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.048334 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-builder-dockercfg-l2p5g-push\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.048379 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.048414 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-builder-dockercfg-l2p5g-pull\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.048437 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-container-storage-root\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.048460 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-node-pullsecrets\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.048482 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-buildcachedir\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.048505 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.048524 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf6b5\" (UniqueName: \"kubernetes.io/projected/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-kube-api-access-tf6b5\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.048546 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-ca-bundles\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.048563 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-buildworkdir\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.048580 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-container-storage-run\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.048607 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-system-configs\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.048627 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-blob-cache\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.048733 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-node-pullsecrets\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.048851 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-container-storage-root\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.049001 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-buildworkdir\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.049276 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-blob-cache\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.049302 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-system-configs\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.049373 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-buildcachedir\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.049384 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-ca-bundles\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.049530 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-container-storage-run\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.049782 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-proxy-ca-bundles\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.056699 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-builder-dockercfg-l2p5g-pull\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.060503 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.065224 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-builder-dockercfg-l2p5g-push\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.070727 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf6b5\" (UniqueName: \"kubernetes.io/projected/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-kube-api-access-tf6b5\") pod \"service-telemetry-framework-index-4-build\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.212529 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:42 crc kubenswrapper[4857]: I0222 00:18:42.455703 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-framework-index-4-build"] Feb 22 00:18:43 crc kubenswrapper[4857]: I0222 00:18:43.465915 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-4-build" event={"ID":"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d","Type":"ContainerStarted","Data":"190ab863c694361087bd8fb1265487fbe9baeb610728a24be632bca6725155f6"} Feb 22 00:18:43 crc kubenswrapper[4857]: I0222 00:18:43.466472 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-4-build" event={"ID":"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d","Type":"ContainerStarted","Data":"4c2359ca1e36d419febe33887608286eaa06156234ec63c81a92cd62b375dea5"} Feb 22 00:18:43 crc kubenswrapper[4857]: E0222 00:18:43.535584 4857 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=8528648506435008715, SKID=, AKID=49:B4:28:90:5C:47:E0:09:27:27:EB:F1:16:CB:58:04:C2:1D:C0:A3 failed: x509: certificate signed by unknown authority" Feb 22 00:18:44 crc kubenswrapper[4857]: I0222 00:18:44.560291 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-framework-index-4-build"] Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.483825 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-framework-index-4-build" podUID="5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" containerName="git-clone" containerID="cri-o://190ab863c694361087bd8fb1265487fbe9baeb610728a24be632bca6725155f6" gracePeriod=30 Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.535477 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-2tmjv"] Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.536414 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-2tmjv" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.543109 4857 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"infrawatch-operators-dockercfg-tpkxj" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.547130 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-2tmjv"] Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.596191 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnzhh\" (UniqueName: \"kubernetes.io/projected/79a1d63a-caef-4011-9e0f-c9a0bf05f04e-kube-api-access-wnzhh\") pod \"infrawatch-operators-2tmjv\" (UID: \"79a1d63a-caef-4011-9e0f-c9a0bf05f04e\") " pod="service-telemetry/infrawatch-operators-2tmjv" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.697750 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnzhh\" (UniqueName: \"kubernetes.io/projected/79a1d63a-caef-4011-9e0f-c9a0bf05f04e-kube-api-access-wnzhh\") pod \"infrawatch-operators-2tmjv\" (UID: \"79a1d63a-caef-4011-9e0f-c9a0bf05f04e\") " pod="service-telemetry/infrawatch-operators-2tmjv" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.717133 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnzhh\" (UniqueName: \"kubernetes.io/projected/79a1d63a-caef-4011-9e0f-c9a0bf05f04e-kube-api-access-wnzhh\") pod \"infrawatch-operators-2tmjv\" (UID: \"79a1d63a-caef-4011-9e0f-c9a0bf05f04e\") " pod="service-telemetry/infrawatch-operators-2tmjv" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.797009 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-framework-index-4-build_5c32c4d2-38c1-42e6-ad56-5d680d7ba14d/git-clone/0.log" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.797095 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.877567 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-2tmjv" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.900169 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-builder-dockercfg-l2p5g-push\") pod \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.900244 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-node-pullsecrets\") pod \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.900292 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-system-configs\") pod \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.900347 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-proxy-ca-bundles\") pod \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.900363 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" (UID: "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.900405 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf6b5\" (UniqueName: \"kubernetes.io/projected/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-kube-api-access-tf6b5\") pod \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.900474 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-blob-cache\") pod \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.900585 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-container-storage-run\") pod \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.900830 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" (UID: "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.900987 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" (UID: "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.901205 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" (UID: "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.901265 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" (UID: "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.901359 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-buildworkdir\") pod \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.901412 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-container-storage-root\") pod \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.901510 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-builder-dockercfg-l2p5g-pull\") pod \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.902130 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-buildcachedir\") pod \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.901681 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" (UID: "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.902000 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" (UID: "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.902287 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-ca-bundles\") pod \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.902369 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" (UID: "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.902550 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-service-telemetry-framework-index-dockercfg-user-build-volume\") pod \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\" (UID: \"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d\") " Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.903281 4857 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.903336 4857 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.903372 4857 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.903398 4857 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.903423 4857 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.903446 4857 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.903467 4857 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.903489 4857 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.903554 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-builder-dockercfg-l2p5g-push" (OuterVolumeSpecName: "builder-dockercfg-l2p5g-push") pod "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" (UID: "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d"). InnerVolumeSpecName "builder-dockercfg-l2p5g-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.903609 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" (UID: "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.903718 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-builder-dockercfg-l2p5g-pull" (OuterVolumeSpecName: "builder-dockercfg-l2p5g-pull") pod "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" (UID: "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d"). InnerVolumeSpecName "builder-dockercfg-l2p5g-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.904013 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-kube-api-access-tf6b5" (OuterVolumeSpecName: "kube-api-access-tf6b5") pod "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" (UID: "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d"). InnerVolumeSpecName "kube-api-access-tf6b5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:18:45 crc kubenswrapper[4857]: I0222 00:18:45.904621 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-service-telemetry-framework-index-dockercfg-user-build-volume" (OuterVolumeSpecName: "service-telemetry-framework-index-dockercfg-user-build-volume") pod "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" (UID: "5c32c4d2-38c1-42e6-ad56-5d680d7ba14d"). InnerVolumeSpecName "service-telemetry-framework-index-dockercfg-user-build-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:18:46 crc kubenswrapper[4857]: I0222 00:18:46.005329 4857 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-l2p5g-pull\" (UniqueName: \"kubernetes.io/secret/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-builder-dockercfg-l2p5g-pull\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:46 crc kubenswrapper[4857]: I0222 00:18:46.005664 4857 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:46 crc kubenswrapper[4857]: I0222 00:18:46.005681 4857 reconciler_common.go:293] "Volume detached for volume \"service-telemetry-framework-index-dockercfg-user-build-volume\" (UniqueName: \"kubernetes.io/secret/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-service-telemetry-framework-index-dockercfg-user-build-volume\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:46 crc kubenswrapper[4857]: I0222 00:18:46.005695 4857 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-l2p5g-push\" (UniqueName: \"kubernetes.io/secret/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-builder-dockercfg-l2p5g-push\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:46 crc kubenswrapper[4857]: I0222 00:18:46.005709 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf6b5\" (UniqueName: \"kubernetes.io/projected/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d-kube-api-access-tf6b5\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:46 crc kubenswrapper[4857]: I0222 00:18:46.045265 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-2tmjv"] Feb 22 00:18:46 crc kubenswrapper[4857]: W0222 00:18:46.047414 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79a1d63a_caef_4011_9e0f_c9a0bf05f04e.slice/crio-8c56723e26331c9982a80adcae771282cfbf32838ffd8f58953a7b25daf2f456 WatchSource:0}: Error finding container 8c56723e26331c9982a80adcae771282cfbf32838ffd8f58953a7b25daf2f456: Status 404 returned error can't find the container with id 8c56723e26331c9982a80adcae771282cfbf32838ffd8f58953a7b25daf2f456 Feb 22 00:18:46 crc kubenswrapper[4857]: E0222 00:18:46.082225 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest" Feb 22 00:18:46 crc kubenswrapper[4857]: E0222 00:18:46.082405 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wnzhh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infrawatch-operators-2tmjv_service-telemetry(79a1d63a-caef-4011-9e0f-c9a0bf05f04e): ErrImagePull: initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" logger="UnhandledError" Feb 22 00:18:46 crc kubenswrapper[4857]: E0222 00:18:46.083878 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown\"" pod="service-telemetry/infrawatch-operators-2tmjv" podUID="79a1d63a-caef-4011-9e0f-c9a0bf05f04e" Feb 22 00:18:46 crc kubenswrapper[4857]: I0222 00:18:46.491541 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-2tmjv" event={"ID":"79a1d63a-caef-4011-9e0f-c9a0bf05f04e","Type":"ContainerStarted","Data":"8c56723e26331c9982a80adcae771282cfbf32838ffd8f58953a7b25daf2f456"} Feb 22 00:18:46 crc kubenswrapper[4857]: E0222 00:18:46.493461 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-2tmjv" podUID="79a1d63a-caef-4011-9e0f-c9a0bf05f04e" Feb 22 00:18:46 crc kubenswrapper[4857]: I0222 00:18:46.495280 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-framework-index-4-build_5c32c4d2-38c1-42e6-ad56-5d680d7ba14d/git-clone/0.log" Feb 22 00:18:46 crc kubenswrapper[4857]: I0222 00:18:46.495560 4857 generic.go:334] "Generic (PLEG): container finished" podID="5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" containerID="190ab863c694361087bd8fb1265487fbe9baeb610728a24be632bca6725155f6" exitCode=1 Feb 22 00:18:46 crc kubenswrapper[4857]: I0222 00:18:46.495620 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-4-build" event={"ID":"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d","Type":"ContainerDied","Data":"190ab863c694361087bd8fb1265487fbe9baeb610728a24be632bca6725155f6"} Feb 22 00:18:46 crc kubenswrapper[4857]: I0222 00:18:46.495674 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-framework-index-4-build" event={"ID":"5c32c4d2-38c1-42e6-ad56-5d680d7ba14d","Type":"ContainerDied","Data":"4c2359ca1e36d419febe33887608286eaa06156234ec63c81a92cd62b375dea5"} Feb 22 00:18:46 crc kubenswrapper[4857]: I0222 00:18:46.495711 4857 scope.go:117] "RemoveContainer" containerID="190ab863c694361087bd8fb1265487fbe9baeb610728a24be632bca6725155f6" Feb 22 00:18:46 crc kubenswrapper[4857]: I0222 00:18:46.495635 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-framework-index-4-build" Feb 22 00:18:46 crc kubenswrapper[4857]: I0222 00:18:46.520570 4857 scope.go:117] "RemoveContainer" containerID="190ab863c694361087bd8fb1265487fbe9baeb610728a24be632bca6725155f6" Feb 22 00:18:46 crc kubenswrapper[4857]: E0222 00:18:46.521259 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"190ab863c694361087bd8fb1265487fbe9baeb610728a24be632bca6725155f6\": container with ID starting with 190ab863c694361087bd8fb1265487fbe9baeb610728a24be632bca6725155f6 not found: ID does not exist" containerID="190ab863c694361087bd8fb1265487fbe9baeb610728a24be632bca6725155f6" Feb 22 00:18:46 crc kubenswrapper[4857]: I0222 00:18:46.521302 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"190ab863c694361087bd8fb1265487fbe9baeb610728a24be632bca6725155f6"} err="failed to get container status \"190ab863c694361087bd8fb1265487fbe9baeb610728a24be632bca6725155f6\": rpc error: code = NotFound desc = could not find container \"190ab863c694361087bd8fb1265487fbe9baeb610728a24be632bca6725155f6\": container with ID starting with 190ab863c694361087bd8fb1265487fbe9baeb610728a24be632bca6725155f6 not found: ID does not exist" Feb 22 00:18:46 crc kubenswrapper[4857]: I0222 00:18:46.541711 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-framework-index-4-build"] Feb 22 00:18:46 crc kubenswrapper[4857]: I0222 00:18:46.547335 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-framework-index-4-build"] Feb 22 00:18:47 crc kubenswrapper[4857]: I0222 00:18:47.089706 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" path="/var/lib/kubelet/pods/5c32c4d2-38c1-42e6-ad56-5d680d7ba14d/volumes" Feb 22 00:18:47 crc kubenswrapper[4857]: E0222 00:18:47.505405 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-2tmjv" podUID="79a1d63a-caef-4011-9e0f-c9a0bf05f04e" Feb 22 00:18:49 crc kubenswrapper[4857]: I0222 00:18:49.924133 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-2tmjv"] Feb 22 00:18:50 crc kubenswrapper[4857]: I0222 00:18:50.721465 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-9lk9c"] Feb 22 00:18:50 crc kubenswrapper[4857]: E0222 00:18:50.721674 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" containerName="git-clone" Feb 22 00:18:50 crc kubenswrapper[4857]: I0222 00:18:50.721686 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" containerName="git-clone" Feb 22 00:18:50 crc kubenswrapper[4857]: I0222 00:18:50.721791 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c32c4d2-38c1-42e6-ad56-5d680d7ba14d" containerName="git-clone" Feb 22 00:18:50 crc kubenswrapper[4857]: I0222 00:18:50.722195 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-9lk9c" Feb 22 00:18:50 crc kubenswrapper[4857]: I0222 00:18:50.730467 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-9lk9c"] Feb 22 00:18:50 crc kubenswrapper[4857]: I0222 00:18:50.763324 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqrvk\" (UniqueName: \"kubernetes.io/projected/57fe7048-f646-407a-a625-0be29af288ff-kube-api-access-bqrvk\") pod \"infrawatch-operators-9lk9c\" (UID: \"57fe7048-f646-407a-a625-0be29af288ff\") " pod="service-telemetry/infrawatch-operators-9lk9c" Feb 22 00:18:50 crc kubenswrapper[4857]: I0222 00:18:50.864637 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqrvk\" (UniqueName: \"kubernetes.io/projected/57fe7048-f646-407a-a625-0be29af288ff-kube-api-access-bqrvk\") pod \"infrawatch-operators-9lk9c\" (UID: \"57fe7048-f646-407a-a625-0be29af288ff\") " pod="service-telemetry/infrawatch-operators-9lk9c" Feb 22 00:18:50 crc kubenswrapper[4857]: I0222 00:18:50.888631 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqrvk\" (UniqueName: \"kubernetes.io/projected/57fe7048-f646-407a-a625-0be29af288ff-kube-api-access-bqrvk\") pod \"infrawatch-operators-9lk9c\" (UID: \"57fe7048-f646-407a-a625-0be29af288ff\") " pod="service-telemetry/infrawatch-operators-9lk9c" Feb 22 00:18:51 crc kubenswrapper[4857]: I0222 00:18:51.037433 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-9lk9c" Feb 22 00:18:51 crc kubenswrapper[4857]: I0222 00:18:51.105094 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-2tmjv" Feb 22 00:18:51 crc kubenswrapper[4857]: I0222 00:18:51.168528 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnzhh\" (UniqueName: \"kubernetes.io/projected/79a1d63a-caef-4011-9e0f-c9a0bf05f04e-kube-api-access-wnzhh\") pod \"79a1d63a-caef-4011-9e0f-c9a0bf05f04e\" (UID: \"79a1d63a-caef-4011-9e0f-c9a0bf05f04e\") " Feb 22 00:18:51 crc kubenswrapper[4857]: I0222 00:18:51.181215 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79a1d63a-caef-4011-9e0f-c9a0bf05f04e-kube-api-access-wnzhh" (OuterVolumeSpecName: "kube-api-access-wnzhh") pod "79a1d63a-caef-4011-9e0f-c9a0bf05f04e" (UID: "79a1d63a-caef-4011-9e0f-c9a0bf05f04e"). InnerVolumeSpecName "kube-api-access-wnzhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:18:51 crc kubenswrapper[4857]: I0222 00:18:51.269467 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnzhh\" (UniqueName: \"kubernetes.io/projected/79a1d63a-caef-4011-9e0f-c9a0bf05f04e-kube-api-access-wnzhh\") on node \"crc\" DevicePath \"\"" Feb 22 00:18:51 crc kubenswrapper[4857]: I0222 00:18:51.476958 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-9lk9c"] Feb 22 00:18:51 crc kubenswrapper[4857]: E0222 00:18:51.508275 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest" Feb 22 00:18:51 crc kubenswrapper[4857]: E0222 00:18:51.508475 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bqrvk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infrawatch-operators-9lk9c_service-telemetry(57fe7048-f646-407a-a625-0be29af288ff): ErrImagePull: initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" logger="UnhandledError" Feb 22 00:18:51 crc kubenswrapper[4857]: E0222 00:18:51.509666 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:18:51 crc kubenswrapper[4857]: I0222 00:18:51.528671 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-2tmjv" event={"ID":"79a1d63a-caef-4011-9e0f-c9a0bf05f04e","Type":"ContainerDied","Data":"8c56723e26331c9982a80adcae771282cfbf32838ffd8f58953a7b25daf2f456"} Feb 22 00:18:51 crc kubenswrapper[4857]: I0222 00:18:51.528694 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-2tmjv" Feb 22 00:18:51 crc kubenswrapper[4857]: I0222 00:18:51.531012 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-9lk9c" event={"ID":"57fe7048-f646-407a-a625-0be29af288ff","Type":"ContainerStarted","Data":"a7340f3eb8a604432f22f1e36969d36d7e89e7b69bb512fde09e27226176a1d8"} Feb 22 00:18:51 crc kubenswrapper[4857]: E0222 00:18:51.532160 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:18:51 crc kubenswrapper[4857]: I0222 00:18:51.599196 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-2tmjv"] Feb 22 00:18:51 crc kubenswrapper[4857]: I0222 00:18:51.604697 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/infrawatch-operators-2tmjv"] Feb 22 00:18:52 crc kubenswrapper[4857]: E0222 00:18:52.539665 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:18:53 crc kubenswrapper[4857]: I0222 00:18:53.084954 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79a1d63a-caef-4011-9e0f-c9a0bf05f04e" path="/var/lib/kubelet/pods/79a1d63a-caef-4011-9e0f-c9a0bf05f04e/volumes" Feb 22 00:19:06 crc kubenswrapper[4857]: E0222 00:19:06.109958 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest" Feb 22 00:19:06 crc kubenswrapper[4857]: E0222 00:19:06.110787 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bqrvk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infrawatch-operators-9lk9c_service-telemetry(57fe7048-f646-407a-a625-0be29af288ff): ErrImagePull: initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" logger="UnhandledError" Feb 22 00:19:06 crc kubenswrapper[4857]: E0222 00:19:06.112082 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:19:09 crc kubenswrapper[4857]: I0222 00:19:09.073016 4857 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 22 00:19:10 crc kubenswrapper[4857]: I0222 00:19:10.428489 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:19:10 crc kubenswrapper[4857]: I0222 00:19:10.428549 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:19:10 crc kubenswrapper[4857]: I0222 00:19:10.428595 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:19:10 crc kubenswrapper[4857]: I0222 00:19:10.429182 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"844b27723312ec8a6f8db792646f1c5f5e12d2eaf2bbdc056197585dfd92820e"} pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 22 00:19:10 crc kubenswrapper[4857]: I0222 00:19:10.429239 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" containerID="cri-o://844b27723312ec8a6f8db792646f1c5f5e12d2eaf2bbdc056197585dfd92820e" gracePeriod=600 Feb 22 00:19:10 crc kubenswrapper[4857]: I0222 00:19:10.648217 4857 generic.go:334] "Generic (PLEG): container finished" podID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerID="844b27723312ec8a6f8db792646f1c5f5e12d2eaf2bbdc056197585dfd92820e" exitCode=0 Feb 22 00:19:10 crc kubenswrapper[4857]: I0222 00:19:10.648265 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" event={"ID":"2ed2fde8-1449-491b-ae21-2fe11ab008ae","Type":"ContainerDied","Data":"844b27723312ec8a6f8db792646f1c5f5e12d2eaf2bbdc056197585dfd92820e"} Feb 22 00:19:10 crc kubenswrapper[4857]: I0222 00:19:10.648307 4857 scope.go:117] "RemoveContainer" containerID="833ca5f17cd1eda283100030bd186ddd6d27df0c724e366f9f63df7686e64ad5" Feb 22 00:19:11 crc kubenswrapper[4857]: I0222 00:19:11.656333 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" event={"ID":"2ed2fde8-1449-491b-ae21-2fe11ab008ae","Type":"ContainerStarted","Data":"d7b4b58f072c91cd783115813cd320a04ba05440e4f3d243be01bcab4bd1e785"} Feb 22 00:19:20 crc kubenswrapper[4857]: E0222 00:19:20.079694 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:19:35 crc kubenswrapper[4857]: E0222 00:19:35.109492 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest" Feb 22 00:19:35 crc kubenswrapper[4857]: E0222 00:19:35.110152 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bqrvk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infrawatch-operators-9lk9c_service-telemetry(57fe7048-f646-407a-a625-0be29af288ff): ErrImagePull: initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" logger="UnhandledError" Feb 22 00:19:35 crc kubenswrapper[4857]: E0222 00:19:35.111712 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:19:50 crc kubenswrapper[4857]: E0222 00:19:50.079611 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:20:05 crc kubenswrapper[4857]: E0222 00:20:05.079641 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:20:17 crc kubenswrapper[4857]: E0222 00:20:17.118085 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest" Feb 22 00:20:17 crc kubenswrapper[4857]: E0222 00:20:17.118780 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bqrvk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infrawatch-operators-9lk9c_service-telemetry(57fe7048-f646-407a-a625-0be29af288ff): ErrImagePull: initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" logger="UnhandledError" Feb 22 00:20:17 crc kubenswrapper[4857]: E0222 00:20:17.120136 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:20:31 crc kubenswrapper[4857]: E0222 00:20:31.078589 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:20:43 crc kubenswrapper[4857]: E0222 00:20:43.081662 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:20:57 crc kubenswrapper[4857]: E0222 00:20:57.079560 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:21:10 crc kubenswrapper[4857]: E0222 00:21:10.079559 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:21:10 crc kubenswrapper[4857]: I0222 00:21:10.428383 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:21:10 crc kubenswrapper[4857]: I0222 00:21:10.428454 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:21:22 crc kubenswrapper[4857]: E0222 00:21:22.080830 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:21:35 crc kubenswrapper[4857]: E0222 00:21:35.080727 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:21:40 crc kubenswrapper[4857]: I0222 00:21:40.428554 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:21:40 crc kubenswrapper[4857]: I0222 00:21:40.428936 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:21:47 crc kubenswrapper[4857]: E0222 00:21:47.116914 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest" Feb 22 00:21:47 crc kubenswrapper[4857]: E0222 00:21:47.117724 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bqrvk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infrawatch-operators-9lk9c_service-telemetry(57fe7048-f646-407a-a625-0be29af288ff): ErrImagePull: initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" logger="UnhandledError" Feb 22 00:21:47 crc kubenswrapper[4857]: E0222 00:21:47.118931 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:21:53 crc kubenswrapper[4857]: I0222 00:21:53.210959 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-75rsl"] Feb 22 00:21:53 crc kubenswrapper[4857]: I0222 00:21:53.212813 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-75rsl" Feb 22 00:21:53 crc kubenswrapper[4857]: I0222 00:21:53.234606 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-75rsl"] Feb 22 00:21:53 crc kubenswrapper[4857]: I0222 00:21:53.279399 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35bf2fbb-a594-427a-abe9-46c9a22ff19b-catalog-content\") pod \"certified-operators-75rsl\" (UID: \"35bf2fbb-a594-427a-abe9-46c9a22ff19b\") " pod="openshift-marketplace/certified-operators-75rsl" Feb 22 00:21:53 crc kubenswrapper[4857]: I0222 00:21:53.279634 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35bf2fbb-a594-427a-abe9-46c9a22ff19b-utilities\") pod \"certified-operators-75rsl\" (UID: \"35bf2fbb-a594-427a-abe9-46c9a22ff19b\") " pod="openshift-marketplace/certified-operators-75rsl" Feb 22 00:21:53 crc kubenswrapper[4857]: I0222 00:21:53.279752 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z42tg\" (UniqueName: \"kubernetes.io/projected/35bf2fbb-a594-427a-abe9-46c9a22ff19b-kube-api-access-z42tg\") pod \"certified-operators-75rsl\" (UID: \"35bf2fbb-a594-427a-abe9-46c9a22ff19b\") " pod="openshift-marketplace/certified-operators-75rsl" Feb 22 00:21:53 crc kubenswrapper[4857]: I0222 00:21:53.380711 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z42tg\" (UniqueName: \"kubernetes.io/projected/35bf2fbb-a594-427a-abe9-46c9a22ff19b-kube-api-access-z42tg\") pod \"certified-operators-75rsl\" (UID: \"35bf2fbb-a594-427a-abe9-46c9a22ff19b\") " pod="openshift-marketplace/certified-operators-75rsl" Feb 22 00:21:53 crc kubenswrapper[4857]: I0222 00:21:53.380812 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35bf2fbb-a594-427a-abe9-46c9a22ff19b-catalog-content\") pod \"certified-operators-75rsl\" (UID: \"35bf2fbb-a594-427a-abe9-46c9a22ff19b\") " pod="openshift-marketplace/certified-operators-75rsl" Feb 22 00:21:53 crc kubenswrapper[4857]: I0222 00:21:53.380843 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35bf2fbb-a594-427a-abe9-46c9a22ff19b-utilities\") pod \"certified-operators-75rsl\" (UID: \"35bf2fbb-a594-427a-abe9-46c9a22ff19b\") " pod="openshift-marketplace/certified-operators-75rsl" Feb 22 00:21:53 crc kubenswrapper[4857]: I0222 00:21:53.381469 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35bf2fbb-a594-427a-abe9-46c9a22ff19b-utilities\") pod \"certified-operators-75rsl\" (UID: \"35bf2fbb-a594-427a-abe9-46c9a22ff19b\") " pod="openshift-marketplace/certified-operators-75rsl" Feb 22 00:21:53 crc kubenswrapper[4857]: I0222 00:21:53.381594 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35bf2fbb-a594-427a-abe9-46c9a22ff19b-catalog-content\") pod \"certified-operators-75rsl\" (UID: \"35bf2fbb-a594-427a-abe9-46c9a22ff19b\") " pod="openshift-marketplace/certified-operators-75rsl" Feb 22 00:21:53 crc kubenswrapper[4857]: I0222 00:21:53.400398 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z42tg\" (UniqueName: \"kubernetes.io/projected/35bf2fbb-a594-427a-abe9-46c9a22ff19b-kube-api-access-z42tg\") pod \"certified-operators-75rsl\" (UID: \"35bf2fbb-a594-427a-abe9-46c9a22ff19b\") " pod="openshift-marketplace/certified-operators-75rsl" Feb 22 00:21:53 crc kubenswrapper[4857]: I0222 00:21:53.530424 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-75rsl" Feb 22 00:21:53 crc kubenswrapper[4857]: I0222 00:21:53.832188 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-75rsl"] Feb 22 00:21:54 crc kubenswrapper[4857]: I0222 00:21:54.668221 4857 generic.go:334] "Generic (PLEG): container finished" podID="35bf2fbb-a594-427a-abe9-46c9a22ff19b" containerID="c180e6e0a54e599696ce8eb2ca202a25274b0294ec0038d72304b63e61a4826c" exitCode=0 Feb 22 00:21:54 crc kubenswrapper[4857]: I0222 00:21:54.668281 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-75rsl" event={"ID":"35bf2fbb-a594-427a-abe9-46c9a22ff19b","Type":"ContainerDied","Data":"c180e6e0a54e599696ce8eb2ca202a25274b0294ec0038d72304b63e61a4826c"} Feb 22 00:21:54 crc kubenswrapper[4857]: I0222 00:21:54.668607 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-75rsl" event={"ID":"35bf2fbb-a594-427a-abe9-46c9a22ff19b","Type":"ContainerStarted","Data":"7501fd7928491f10cb8ab48f650aaf2459de585b4955073f7190ae945f05ccd2"} Feb 22 00:21:55 crc kubenswrapper[4857]: I0222 00:21:55.675181 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-75rsl" event={"ID":"35bf2fbb-a594-427a-abe9-46c9a22ff19b","Type":"ContainerStarted","Data":"012c442310468a8b8a10deaf7eb442305da5337e8dfefefd379b536a89595033"} Feb 22 00:21:56 crc kubenswrapper[4857]: I0222 00:21:56.683776 4857 generic.go:334] "Generic (PLEG): container finished" podID="35bf2fbb-a594-427a-abe9-46c9a22ff19b" containerID="012c442310468a8b8a10deaf7eb442305da5337e8dfefefd379b536a89595033" exitCode=0 Feb 22 00:21:56 crc kubenswrapper[4857]: I0222 00:21:56.683827 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-75rsl" event={"ID":"35bf2fbb-a594-427a-abe9-46c9a22ff19b","Type":"ContainerDied","Data":"012c442310468a8b8a10deaf7eb442305da5337e8dfefefd379b536a89595033"} Feb 22 00:21:56 crc kubenswrapper[4857]: I0222 00:21:56.687438 4857 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 22 00:21:57 crc kubenswrapper[4857]: I0222 00:21:57.695844 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-75rsl" event={"ID":"35bf2fbb-a594-427a-abe9-46c9a22ff19b","Type":"ContainerStarted","Data":"7d76ebfd856d81093c3b3515491bceaca67682b3cca7e4dea80bc2a7baf7264f"} Feb 22 00:21:57 crc kubenswrapper[4857]: I0222 00:21:57.722839 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-75rsl" podStartSLOduration=2.265489357 podStartE2EDuration="4.722815159s" podCreationTimestamp="2026-02-22 00:21:53 +0000 UTC" firstStartedPulling="2026-02-22 00:21:54.670408035 +0000 UTC m=+922.309137288" lastFinishedPulling="2026-02-22 00:21:57.127733837 +0000 UTC m=+924.766463090" observedRunningTime="2026-02-22 00:21:57.717807725 +0000 UTC m=+925.356536988" watchObservedRunningTime="2026-02-22 00:21:57.722815159 +0000 UTC m=+925.361544422" Feb 22 00:21:59 crc kubenswrapper[4857]: I0222 00:21:59.601494 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x29rh"] Feb 22 00:21:59 crc kubenswrapper[4857]: I0222 00:21:59.603567 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x29rh" Feb 22 00:21:59 crc kubenswrapper[4857]: I0222 00:21:59.610691 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x29rh"] Feb 22 00:21:59 crc kubenswrapper[4857]: I0222 00:21:59.799309 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xljv6\" (UniqueName: \"kubernetes.io/projected/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0-kube-api-access-xljv6\") pod \"redhat-operators-x29rh\" (UID: \"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0\") " pod="openshift-marketplace/redhat-operators-x29rh" Feb 22 00:21:59 crc kubenswrapper[4857]: I0222 00:21:59.799527 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0-utilities\") pod \"redhat-operators-x29rh\" (UID: \"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0\") " pod="openshift-marketplace/redhat-operators-x29rh" Feb 22 00:21:59 crc kubenswrapper[4857]: I0222 00:21:59.799606 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0-catalog-content\") pod \"redhat-operators-x29rh\" (UID: \"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0\") " pod="openshift-marketplace/redhat-operators-x29rh" Feb 22 00:21:59 crc kubenswrapper[4857]: I0222 00:21:59.900433 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0-utilities\") pod \"redhat-operators-x29rh\" (UID: \"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0\") " pod="openshift-marketplace/redhat-operators-x29rh" Feb 22 00:21:59 crc kubenswrapper[4857]: I0222 00:21:59.900737 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0-catalog-content\") pod \"redhat-operators-x29rh\" (UID: \"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0\") " pod="openshift-marketplace/redhat-operators-x29rh" Feb 22 00:21:59 crc kubenswrapper[4857]: I0222 00:21:59.900929 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xljv6\" (UniqueName: \"kubernetes.io/projected/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0-kube-api-access-xljv6\") pod \"redhat-operators-x29rh\" (UID: \"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0\") " pod="openshift-marketplace/redhat-operators-x29rh" Feb 22 00:21:59 crc kubenswrapper[4857]: I0222 00:21:59.901339 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0-catalog-content\") pod \"redhat-operators-x29rh\" (UID: \"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0\") " pod="openshift-marketplace/redhat-operators-x29rh" Feb 22 00:21:59 crc kubenswrapper[4857]: I0222 00:21:59.901340 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0-utilities\") pod \"redhat-operators-x29rh\" (UID: \"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0\") " pod="openshift-marketplace/redhat-operators-x29rh" Feb 22 00:21:59 crc kubenswrapper[4857]: I0222 00:21:59.926738 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xljv6\" (UniqueName: \"kubernetes.io/projected/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0-kube-api-access-xljv6\") pod \"redhat-operators-x29rh\" (UID: \"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0\") " pod="openshift-marketplace/redhat-operators-x29rh" Feb 22 00:21:59 crc kubenswrapper[4857]: I0222 00:21:59.962447 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x29rh" Feb 22 00:22:00 crc kubenswrapper[4857]: E0222 00:22:00.079380 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:22:00 crc kubenswrapper[4857]: I0222 00:22:00.152559 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x29rh"] Feb 22 00:22:00 crc kubenswrapper[4857]: W0222 00:22:00.156424 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a2e5c63_36aa_406f_b5fb_3a62efd94bc0.slice/crio-b45239ad7b0daae9b92b7277d5c0379674f4fd36d155e74ed165d0014325929a WatchSource:0}: Error finding container b45239ad7b0daae9b92b7277d5c0379674f4fd36d155e74ed165d0014325929a: Status 404 returned error can't find the container with id b45239ad7b0daae9b92b7277d5c0379674f4fd36d155e74ed165d0014325929a Feb 22 00:22:00 crc kubenswrapper[4857]: I0222 00:22:00.714199 4857 generic.go:334] "Generic (PLEG): container finished" podID="3a2e5c63-36aa-406f-b5fb-3a62efd94bc0" containerID="65c0a561f354d62a0a69c657478f52dce42c27f46c62fef722d77f177fd2d7ba" exitCode=0 Feb 22 00:22:00 crc kubenswrapper[4857]: I0222 00:22:00.714255 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x29rh" event={"ID":"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0","Type":"ContainerDied","Data":"65c0a561f354d62a0a69c657478f52dce42c27f46c62fef722d77f177fd2d7ba"} Feb 22 00:22:00 crc kubenswrapper[4857]: I0222 00:22:00.714534 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x29rh" event={"ID":"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0","Type":"ContainerStarted","Data":"b45239ad7b0daae9b92b7277d5c0379674f4fd36d155e74ed165d0014325929a"} Feb 22 00:22:01 crc kubenswrapper[4857]: I0222 00:22:01.721889 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x29rh" event={"ID":"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0","Type":"ContainerStarted","Data":"d9f5f841ae883f4de112424162e2288ad80d6380f105259c27a9700140576380"} Feb 22 00:22:02 crc kubenswrapper[4857]: I0222 00:22:02.731338 4857 generic.go:334] "Generic (PLEG): container finished" podID="3a2e5c63-36aa-406f-b5fb-3a62efd94bc0" containerID="d9f5f841ae883f4de112424162e2288ad80d6380f105259c27a9700140576380" exitCode=0 Feb 22 00:22:02 crc kubenswrapper[4857]: I0222 00:22:02.731409 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x29rh" event={"ID":"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0","Type":"ContainerDied","Data":"d9f5f841ae883f4de112424162e2288ad80d6380f105259c27a9700140576380"} Feb 22 00:22:03 crc kubenswrapper[4857]: I0222 00:22:03.531201 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-75rsl" Feb 22 00:22:03 crc kubenswrapper[4857]: I0222 00:22:03.531970 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-75rsl" Feb 22 00:22:03 crc kubenswrapper[4857]: I0222 00:22:03.569909 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-75rsl" Feb 22 00:22:03 crc kubenswrapper[4857]: I0222 00:22:03.739779 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x29rh" event={"ID":"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0","Type":"ContainerStarted","Data":"30344a07c6dd68e5604e0209e375b45d81a450e2e796ca0bef34f0b4736fd774"} Feb 22 00:22:03 crc kubenswrapper[4857]: I0222 00:22:03.777910 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-75rsl" Feb 22 00:22:03 crc kubenswrapper[4857]: I0222 00:22:03.796983 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x29rh" podStartSLOduration=2.384556667 podStartE2EDuration="4.796960546s" podCreationTimestamp="2026-02-22 00:21:59 +0000 UTC" firstStartedPulling="2026-02-22 00:22:00.71567069 +0000 UTC m=+928.354399943" lastFinishedPulling="2026-02-22 00:22:03.128074569 +0000 UTC m=+930.766803822" observedRunningTime="2026-02-22 00:22:03.76042401 +0000 UTC m=+931.399153273" watchObservedRunningTime="2026-02-22 00:22:03.796960546 +0000 UTC m=+931.435689799" Feb 22 00:22:05 crc kubenswrapper[4857]: I0222 00:22:05.987803 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-75rsl"] Feb 22 00:22:06 crc kubenswrapper[4857]: I0222 00:22:06.755367 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-75rsl" podUID="35bf2fbb-a594-427a-abe9-46c9a22ff19b" containerName="registry-server" containerID="cri-o://7d76ebfd856d81093c3b3515491bceaca67682b3cca7e4dea80bc2a7baf7264f" gracePeriod=2 Feb 22 00:22:09 crc kubenswrapper[4857]: I0222 00:22:09.778396 4857 generic.go:334] "Generic (PLEG): container finished" podID="35bf2fbb-a594-427a-abe9-46c9a22ff19b" containerID="7d76ebfd856d81093c3b3515491bceaca67682b3cca7e4dea80bc2a7baf7264f" exitCode=0 Feb 22 00:22:09 crc kubenswrapper[4857]: I0222 00:22:09.778482 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-75rsl" event={"ID":"35bf2fbb-a594-427a-abe9-46c9a22ff19b","Type":"ContainerDied","Data":"7d76ebfd856d81093c3b3515491bceaca67682b3cca7e4dea80bc2a7baf7264f"} Feb 22 00:22:09 crc kubenswrapper[4857]: I0222 00:22:09.962954 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x29rh" Feb 22 00:22:09 crc kubenswrapper[4857]: I0222 00:22:09.963004 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x29rh" Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.018769 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x29rh" Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.428267 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.428343 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.428396 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.429158 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d7b4b58f072c91cd783115813cd320a04ba05440e4f3d243be01bcab4bd1e785"} pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.429221 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" containerID="cri-o://d7b4b58f072c91cd783115813cd320a04ba05440e4f3d243be01bcab4bd1e785" gracePeriod=600 Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.780605 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-75rsl" Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.787518 4857 generic.go:334] "Generic (PLEG): container finished" podID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerID="d7b4b58f072c91cd783115813cd320a04ba05440e4f3d243be01bcab4bd1e785" exitCode=0 Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.787599 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" event={"ID":"2ed2fde8-1449-491b-ae21-2fe11ab008ae","Type":"ContainerDied","Data":"d7b4b58f072c91cd783115813cd320a04ba05440e4f3d243be01bcab4bd1e785"} Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.787652 4857 scope.go:117] "RemoveContainer" containerID="844b27723312ec8a6f8db792646f1c5f5e12d2eaf2bbdc056197585dfd92820e" Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.791191 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-75rsl" event={"ID":"35bf2fbb-a594-427a-abe9-46c9a22ff19b","Type":"ContainerDied","Data":"7501fd7928491f10cb8ab48f650aaf2459de585b4955073f7190ae945f05ccd2"} Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.791234 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-75rsl" Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.837867 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x29rh" Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.874296 4857 scope.go:117] "RemoveContainer" containerID="7d76ebfd856d81093c3b3515491bceaca67682b3cca7e4dea80bc2a7baf7264f" Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.881505 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35bf2fbb-a594-427a-abe9-46c9a22ff19b-utilities\") pod \"35bf2fbb-a594-427a-abe9-46c9a22ff19b\" (UID: \"35bf2fbb-a594-427a-abe9-46c9a22ff19b\") " Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.881573 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35bf2fbb-a594-427a-abe9-46c9a22ff19b-catalog-content\") pod \"35bf2fbb-a594-427a-abe9-46c9a22ff19b\" (UID: \"35bf2fbb-a594-427a-abe9-46c9a22ff19b\") " Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.881602 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z42tg\" (UniqueName: \"kubernetes.io/projected/35bf2fbb-a594-427a-abe9-46c9a22ff19b-kube-api-access-z42tg\") pod \"35bf2fbb-a594-427a-abe9-46c9a22ff19b\" (UID: \"35bf2fbb-a594-427a-abe9-46c9a22ff19b\") " Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.882649 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35bf2fbb-a594-427a-abe9-46c9a22ff19b-utilities" (OuterVolumeSpecName: "utilities") pod "35bf2fbb-a594-427a-abe9-46c9a22ff19b" (UID: "35bf2fbb-a594-427a-abe9-46c9a22ff19b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.889256 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35bf2fbb-a594-427a-abe9-46c9a22ff19b-kube-api-access-z42tg" (OuterVolumeSpecName: "kube-api-access-z42tg") pod "35bf2fbb-a594-427a-abe9-46c9a22ff19b" (UID: "35bf2fbb-a594-427a-abe9-46c9a22ff19b"). InnerVolumeSpecName "kube-api-access-z42tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.891841 4857 scope.go:117] "RemoveContainer" containerID="012c442310468a8b8a10deaf7eb442305da5337e8dfefefd379b536a89595033" Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.923103 4857 scope.go:117] "RemoveContainer" containerID="c180e6e0a54e599696ce8eb2ca202a25274b0294ec0038d72304b63e61a4826c" Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.929450 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35bf2fbb-a594-427a-abe9-46c9a22ff19b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35bf2fbb-a594-427a-abe9-46c9a22ff19b" (UID: "35bf2fbb-a594-427a-abe9-46c9a22ff19b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.983921 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35bf2fbb-a594-427a-abe9-46c9a22ff19b-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.983960 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z42tg\" (UniqueName: \"kubernetes.io/projected/35bf2fbb-a594-427a-abe9-46c9a22ff19b-kube-api-access-z42tg\") on node \"crc\" DevicePath \"\"" Feb 22 00:22:10 crc kubenswrapper[4857]: I0222 00:22:10.983975 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35bf2fbb-a594-427a-abe9-46c9a22ff19b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:22:11 crc kubenswrapper[4857]: I0222 00:22:11.137157 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-75rsl"] Feb 22 00:22:11 crc kubenswrapper[4857]: I0222 00:22:11.143203 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-75rsl"] Feb 22 00:22:11 crc kubenswrapper[4857]: I0222 00:22:11.802538 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" event={"ID":"2ed2fde8-1449-491b-ae21-2fe11ab008ae","Type":"ContainerStarted","Data":"a2bc37e253a36b33e8ab447c1e8c471a24067bf7db0e01c4c28c06ec0eef8f2c"} Feb 22 00:22:12 crc kubenswrapper[4857]: E0222 00:22:12.079198 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.086455 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35bf2fbb-a594-427a-abe9-46c9a22ff19b" path="/var/lib/kubelet/pods/35bf2fbb-a594-427a-abe9-46c9a22ff19b/volumes" Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.189590 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x29rh"] Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.189834 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x29rh" podUID="3a2e5c63-36aa-406f-b5fb-3a62efd94bc0" containerName="registry-server" containerID="cri-o://30344a07c6dd68e5604e0209e375b45d81a450e2e796ca0bef34f0b4736fd774" gracePeriod=2 Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.571595 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x29rh" Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.712911 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0-catalog-content\") pod \"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0\" (UID: \"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0\") " Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.712993 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xljv6\" (UniqueName: \"kubernetes.io/projected/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0-kube-api-access-xljv6\") pod \"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0\" (UID: \"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0\") " Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.713057 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0-utilities\") pod \"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0\" (UID: \"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0\") " Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.714189 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0-utilities" (OuterVolumeSpecName: "utilities") pod "3a2e5c63-36aa-406f-b5fb-3a62efd94bc0" (UID: "3a2e5c63-36aa-406f-b5fb-3a62efd94bc0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.722522 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0-kube-api-access-xljv6" (OuterVolumeSpecName: "kube-api-access-xljv6") pod "3a2e5c63-36aa-406f-b5fb-3a62efd94bc0" (UID: "3a2e5c63-36aa-406f-b5fb-3a62efd94bc0"). InnerVolumeSpecName "kube-api-access-xljv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.814405 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xljv6\" (UniqueName: \"kubernetes.io/projected/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0-kube-api-access-xljv6\") on node \"crc\" DevicePath \"\"" Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.814452 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.818947 4857 generic.go:334] "Generic (PLEG): container finished" podID="3a2e5c63-36aa-406f-b5fb-3a62efd94bc0" containerID="30344a07c6dd68e5604e0209e375b45d81a450e2e796ca0bef34f0b4736fd774" exitCode=0 Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.819049 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x29rh" event={"ID":"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0","Type":"ContainerDied","Data":"30344a07c6dd68e5604e0209e375b45d81a450e2e796ca0bef34f0b4736fd774"} Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.819074 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x29rh" Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.819103 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x29rh" event={"ID":"3a2e5c63-36aa-406f-b5fb-3a62efd94bc0","Type":"ContainerDied","Data":"b45239ad7b0daae9b92b7277d5c0379674f4fd36d155e74ed165d0014325929a"} Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.819127 4857 scope.go:117] "RemoveContainer" containerID="30344a07c6dd68e5604e0209e375b45d81a450e2e796ca0bef34f0b4736fd774" Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.836915 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a2e5c63-36aa-406f-b5fb-3a62efd94bc0" (UID: "3a2e5c63-36aa-406f-b5fb-3a62efd94bc0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.838515 4857 scope.go:117] "RemoveContainer" containerID="d9f5f841ae883f4de112424162e2288ad80d6380f105259c27a9700140576380" Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.859631 4857 scope.go:117] "RemoveContainer" containerID="65c0a561f354d62a0a69c657478f52dce42c27f46c62fef722d77f177fd2d7ba" Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.881603 4857 scope.go:117] "RemoveContainer" containerID="30344a07c6dd68e5604e0209e375b45d81a450e2e796ca0bef34f0b4736fd774" Feb 22 00:22:13 crc kubenswrapper[4857]: E0222 00:22:13.882020 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30344a07c6dd68e5604e0209e375b45d81a450e2e796ca0bef34f0b4736fd774\": container with ID starting with 30344a07c6dd68e5604e0209e375b45d81a450e2e796ca0bef34f0b4736fd774 not found: ID does not exist" containerID="30344a07c6dd68e5604e0209e375b45d81a450e2e796ca0bef34f0b4736fd774" Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.882095 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30344a07c6dd68e5604e0209e375b45d81a450e2e796ca0bef34f0b4736fd774"} err="failed to get container status \"30344a07c6dd68e5604e0209e375b45d81a450e2e796ca0bef34f0b4736fd774\": rpc error: code = NotFound desc = could not find container \"30344a07c6dd68e5604e0209e375b45d81a450e2e796ca0bef34f0b4736fd774\": container with ID starting with 30344a07c6dd68e5604e0209e375b45d81a450e2e796ca0bef34f0b4736fd774 not found: ID does not exist" Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.882132 4857 scope.go:117] "RemoveContainer" containerID="d9f5f841ae883f4de112424162e2288ad80d6380f105259c27a9700140576380" Feb 22 00:22:13 crc kubenswrapper[4857]: E0222 00:22:13.883155 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9f5f841ae883f4de112424162e2288ad80d6380f105259c27a9700140576380\": container with ID starting with d9f5f841ae883f4de112424162e2288ad80d6380f105259c27a9700140576380 not found: ID does not exist" containerID="d9f5f841ae883f4de112424162e2288ad80d6380f105259c27a9700140576380" Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.883191 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9f5f841ae883f4de112424162e2288ad80d6380f105259c27a9700140576380"} err="failed to get container status \"d9f5f841ae883f4de112424162e2288ad80d6380f105259c27a9700140576380\": rpc error: code = NotFound desc = could not find container \"d9f5f841ae883f4de112424162e2288ad80d6380f105259c27a9700140576380\": container with ID starting with d9f5f841ae883f4de112424162e2288ad80d6380f105259c27a9700140576380 not found: ID does not exist" Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.883214 4857 scope.go:117] "RemoveContainer" containerID="65c0a561f354d62a0a69c657478f52dce42c27f46c62fef722d77f177fd2d7ba" Feb 22 00:22:13 crc kubenswrapper[4857]: E0222 00:22:13.883672 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65c0a561f354d62a0a69c657478f52dce42c27f46c62fef722d77f177fd2d7ba\": container with ID starting with 65c0a561f354d62a0a69c657478f52dce42c27f46c62fef722d77f177fd2d7ba not found: ID does not exist" containerID="65c0a561f354d62a0a69c657478f52dce42c27f46c62fef722d77f177fd2d7ba" Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.883707 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65c0a561f354d62a0a69c657478f52dce42c27f46c62fef722d77f177fd2d7ba"} err="failed to get container status \"65c0a561f354d62a0a69c657478f52dce42c27f46c62fef722d77f177fd2d7ba\": rpc error: code = NotFound desc = could not find container \"65c0a561f354d62a0a69c657478f52dce42c27f46c62fef722d77f177fd2d7ba\": container with ID starting with 65c0a561f354d62a0a69c657478f52dce42c27f46c62fef722d77f177fd2d7ba not found: ID does not exist" Feb 22 00:22:13 crc kubenswrapper[4857]: I0222 00:22:13.915758 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:22:14 crc kubenswrapper[4857]: I0222 00:22:14.148010 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x29rh"] Feb 22 00:22:14 crc kubenswrapper[4857]: I0222 00:22:14.152989 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x29rh"] Feb 22 00:22:15 crc kubenswrapper[4857]: I0222 00:22:15.085474 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a2e5c63-36aa-406f-b5fb-3a62efd94bc0" path="/var/lib/kubelet/pods/3a2e5c63-36aa-406f-b5fb-3a62efd94bc0/volumes" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.599907 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8vwcq"] Feb 22 00:22:17 crc kubenswrapper[4857]: E0222 00:22:17.600191 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a2e5c63-36aa-406f-b5fb-3a62efd94bc0" containerName="extract-utilities" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.600206 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a2e5c63-36aa-406f-b5fb-3a62efd94bc0" containerName="extract-utilities" Feb 22 00:22:17 crc kubenswrapper[4857]: E0222 00:22:17.600229 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a2e5c63-36aa-406f-b5fb-3a62efd94bc0" containerName="extract-content" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.600238 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a2e5c63-36aa-406f-b5fb-3a62efd94bc0" containerName="extract-content" Feb 22 00:22:17 crc kubenswrapper[4857]: E0222 00:22:17.600255 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35bf2fbb-a594-427a-abe9-46c9a22ff19b" containerName="registry-server" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.600265 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="35bf2fbb-a594-427a-abe9-46c9a22ff19b" containerName="registry-server" Feb 22 00:22:17 crc kubenswrapper[4857]: E0222 00:22:17.600282 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a2e5c63-36aa-406f-b5fb-3a62efd94bc0" containerName="registry-server" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.600291 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a2e5c63-36aa-406f-b5fb-3a62efd94bc0" containerName="registry-server" Feb 22 00:22:17 crc kubenswrapper[4857]: E0222 00:22:17.600303 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35bf2fbb-a594-427a-abe9-46c9a22ff19b" containerName="extract-utilities" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.600312 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="35bf2fbb-a594-427a-abe9-46c9a22ff19b" containerName="extract-utilities" Feb 22 00:22:17 crc kubenswrapper[4857]: E0222 00:22:17.600320 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35bf2fbb-a594-427a-abe9-46c9a22ff19b" containerName="extract-content" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.600328 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="35bf2fbb-a594-427a-abe9-46c9a22ff19b" containerName="extract-content" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.600454 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="35bf2fbb-a594-427a-abe9-46c9a22ff19b" containerName="registry-server" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.600475 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a2e5c63-36aa-406f-b5fb-3a62efd94bc0" containerName="registry-server" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.601421 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8vwcq" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.613734 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8vwcq"] Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.765554 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8bcd34c-0ae9-41cf-b579-83717bbe623e-catalog-content\") pod \"community-operators-8vwcq\" (UID: \"f8bcd34c-0ae9-41cf-b579-83717bbe623e\") " pod="openshift-marketplace/community-operators-8vwcq" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.766022 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98n2b\" (UniqueName: \"kubernetes.io/projected/f8bcd34c-0ae9-41cf-b579-83717bbe623e-kube-api-access-98n2b\") pod \"community-operators-8vwcq\" (UID: \"f8bcd34c-0ae9-41cf-b579-83717bbe623e\") " pod="openshift-marketplace/community-operators-8vwcq" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.766206 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8bcd34c-0ae9-41cf-b579-83717bbe623e-utilities\") pod \"community-operators-8vwcq\" (UID: \"f8bcd34c-0ae9-41cf-b579-83717bbe623e\") " pod="openshift-marketplace/community-operators-8vwcq" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.868747 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98n2b\" (UniqueName: \"kubernetes.io/projected/f8bcd34c-0ae9-41cf-b579-83717bbe623e-kube-api-access-98n2b\") pod \"community-operators-8vwcq\" (UID: \"f8bcd34c-0ae9-41cf-b579-83717bbe623e\") " pod="openshift-marketplace/community-operators-8vwcq" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.868904 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8bcd34c-0ae9-41cf-b579-83717bbe623e-utilities\") pod \"community-operators-8vwcq\" (UID: \"f8bcd34c-0ae9-41cf-b579-83717bbe623e\") " pod="openshift-marketplace/community-operators-8vwcq" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.868968 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8bcd34c-0ae9-41cf-b579-83717bbe623e-catalog-content\") pod \"community-operators-8vwcq\" (UID: \"f8bcd34c-0ae9-41cf-b579-83717bbe623e\") " pod="openshift-marketplace/community-operators-8vwcq" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.869529 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8bcd34c-0ae9-41cf-b579-83717bbe623e-catalog-content\") pod \"community-operators-8vwcq\" (UID: \"f8bcd34c-0ae9-41cf-b579-83717bbe623e\") " pod="openshift-marketplace/community-operators-8vwcq" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.869736 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8bcd34c-0ae9-41cf-b579-83717bbe623e-utilities\") pod \"community-operators-8vwcq\" (UID: \"f8bcd34c-0ae9-41cf-b579-83717bbe623e\") " pod="openshift-marketplace/community-operators-8vwcq" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.889470 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98n2b\" (UniqueName: \"kubernetes.io/projected/f8bcd34c-0ae9-41cf-b579-83717bbe623e-kube-api-access-98n2b\") pod \"community-operators-8vwcq\" (UID: \"f8bcd34c-0ae9-41cf-b579-83717bbe623e\") " pod="openshift-marketplace/community-operators-8vwcq" Feb 22 00:22:17 crc kubenswrapper[4857]: I0222 00:22:17.923609 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8vwcq" Feb 22 00:22:18 crc kubenswrapper[4857]: I0222 00:22:18.383875 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8vwcq"] Feb 22 00:22:18 crc kubenswrapper[4857]: I0222 00:22:18.853465 4857 generic.go:334] "Generic (PLEG): container finished" podID="f8bcd34c-0ae9-41cf-b579-83717bbe623e" containerID="1160695d13a934e09ea55f880b0a0dafbf9abb0e63c78ca8248d931e89187aee" exitCode=0 Feb 22 00:22:18 crc kubenswrapper[4857]: I0222 00:22:18.853574 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8vwcq" event={"ID":"f8bcd34c-0ae9-41cf-b579-83717bbe623e","Type":"ContainerDied","Data":"1160695d13a934e09ea55f880b0a0dafbf9abb0e63c78ca8248d931e89187aee"} Feb 22 00:22:18 crc kubenswrapper[4857]: I0222 00:22:18.853764 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8vwcq" event={"ID":"f8bcd34c-0ae9-41cf-b579-83717bbe623e","Type":"ContainerStarted","Data":"52fa8cb0fb151f93a4c37d22a432e3cb8627ecdffec6bf8fc8dae84ffba24819"} Feb 22 00:22:19 crc kubenswrapper[4857]: I0222 00:22:19.867644 4857 generic.go:334] "Generic (PLEG): container finished" podID="f8bcd34c-0ae9-41cf-b579-83717bbe623e" containerID="762157d0d8af1c189d65e10c5c3fdfbff510ca0be9532852d55617eefcf35ac0" exitCode=0 Feb 22 00:22:19 crc kubenswrapper[4857]: I0222 00:22:19.867726 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8vwcq" event={"ID":"f8bcd34c-0ae9-41cf-b579-83717bbe623e","Type":"ContainerDied","Data":"762157d0d8af1c189d65e10c5c3fdfbff510ca0be9532852d55617eefcf35ac0"} Feb 22 00:22:20 crc kubenswrapper[4857]: I0222 00:22:20.875415 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8vwcq" event={"ID":"f8bcd34c-0ae9-41cf-b579-83717bbe623e","Type":"ContainerStarted","Data":"300e129cdcf8b8d7284ead7f99b6e1b5e097938101e0073f735b5f97f1d6e063"} Feb 22 00:22:20 crc kubenswrapper[4857]: I0222 00:22:20.894307 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8vwcq" podStartSLOduration=2.480539615 podStartE2EDuration="3.894292061s" podCreationTimestamp="2026-02-22 00:22:17 +0000 UTC" firstStartedPulling="2026-02-22 00:22:18.85457168 +0000 UTC m=+946.493300983" lastFinishedPulling="2026-02-22 00:22:20.268324176 +0000 UTC m=+947.907053429" observedRunningTime="2026-02-22 00:22:20.889316197 +0000 UTC m=+948.528045450" watchObservedRunningTime="2026-02-22 00:22:20.894292061 +0000 UTC m=+948.533021314" Feb 22 00:22:27 crc kubenswrapper[4857]: E0222 00:22:27.080021 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:22:27 crc kubenswrapper[4857]: I0222 00:22:27.924676 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8vwcq" Feb 22 00:22:27 crc kubenswrapper[4857]: I0222 00:22:27.924764 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8vwcq" Feb 22 00:22:27 crc kubenswrapper[4857]: I0222 00:22:27.982952 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8vwcq" Feb 22 00:22:28 crc kubenswrapper[4857]: I0222 00:22:28.982973 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8vwcq" Feb 22 00:22:29 crc kubenswrapper[4857]: I0222 00:22:29.031730 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8vwcq"] Feb 22 00:22:30 crc kubenswrapper[4857]: I0222 00:22:30.936459 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8vwcq" podUID="f8bcd34c-0ae9-41cf-b579-83717bbe623e" containerName="registry-server" containerID="cri-o://300e129cdcf8b8d7284ead7f99b6e1b5e097938101e0073f735b5f97f1d6e063" gracePeriod=2 Feb 22 00:22:31 crc kubenswrapper[4857]: I0222 00:22:31.287782 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8vwcq" Feb 22 00:22:31 crc kubenswrapper[4857]: I0222 00:22:31.457685 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98n2b\" (UniqueName: \"kubernetes.io/projected/f8bcd34c-0ae9-41cf-b579-83717bbe623e-kube-api-access-98n2b\") pod \"f8bcd34c-0ae9-41cf-b579-83717bbe623e\" (UID: \"f8bcd34c-0ae9-41cf-b579-83717bbe623e\") " Feb 22 00:22:31 crc kubenswrapper[4857]: I0222 00:22:31.457736 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8bcd34c-0ae9-41cf-b579-83717bbe623e-catalog-content\") pod \"f8bcd34c-0ae9-41cf-b579-83717bbe623e\" (UID: \"f8bcd34c-0ae9-41cf-b579-83717bbe623e\") " Feb 22 00:22:31 crc kubenswrapper[4857]: I0222 00:22:31.457845 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8bcd34c-0ae9-41cf-b579-83717bbe623e-utilities\") pod \"f8bcd34c-0ae9-41cf-b579-83717bbe623e\" (UID: \"f8bcd34c-0ae9-41cf-b579-83717bbe623e\") " Feb 22 00:22:31 crc kubenswrapper[4857]: I0222 00:22:31.458939 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8bcd34c-0ae9-41cf-b579-83717bbe623e-utilities" (OuterVolumeSpecName: "utilities") pod "f8bcd34c-0ae9-41cf-b579-83717bbe623e" (UID: "f8bcd34c-0ae9-41cf-b579-83717bbe623e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:22:31 crc kubenswrapper[4857]: I0222 00:22:31.463252 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8bcd34c-0ae9-41cf-b579-83717bbe623e-kube-api-access-98n2b" (OuterVolumeSpecName: "kube-api-access-98n2b") pod "f8bcd34c-0ae9-41cf-b579-83717bbe623e" (UID: "f8bcd34c-0ae9-41cf-b579-83717bbe623e"). InnerVolumeSpecName "kube-api-access-98n2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:22:31 crc kubenswrapper[4857]: I0222 00:22:31.509255 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8bcd34c-0ae9-41cf-b579-83717bbe623e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8bcd34c-0ae9-41cf-b579-83717bbe623e" (UID: "f8bcd34c-0ae9-41cf-b579-83717bbe623e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:22:31 crc kubenswrapper[4857]: I0222 00:22:31.559400 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98n2b\" (UniqueName: \"kubernetes.io/projected/f8bcd34c-0ae9-41cf-b579-83717bbe623e-kube-api-access-98n2b\") on node \"crc\" DevicePath \"\"" Feb 22 00:22:31 crc kubenswrapper[4857]: I0222 00:22:31.559649 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8bcd34c-0ae9-41cf-b579-83717bbe623e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:22:31 crc kubenswrapper[4857]: I0222 00:22:31.559731 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8bcd34c-0ae9-41cf-b579-83717bbe623e-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:22:31 crc kubenswrapper[4857]: I0222 00:22:31.949566 4857 generic.go:334] "Generic (PLEG): container finished" podID="f8bcd34c-0ae9-41cf-b579-83717bbe623e" containerID="300e129cdcf8b8d7284ead7f99b6e1b5e097938101e0073f735b5f97f1d6e063" exitCode=0 Feb 22 00:22:31 crc kubenswrapper[4857]: I0222 00:22:31.949618 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8vwcq" event={"ID":"f8bcd34c-0ae9-41cf-b579-83717bbe623e","Type":"ContainerDied","Data":"300e129cdcf8b8d7284ead7f99b6e1b5e097938101e0073f735b5f97f1d6e063"} Feb 22 00:22:31 crc kubenswrapper[4857]: I0222 00:22:31.949649 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8vwcq" event={"ID":"f8bcd34c-0ae9-41cf-b579-83717bbe623e","Type":"ContainerDied","Data":"52fa8cb0fb151f93a4c37d22a432e3cb8627ecdffec6bf8fc8dae84ffba24819"} Feb 22 00:22:31 crc kubenswrapper[4857]: I0222 00:22:31.949673 4857 scope.go:117] "RemoveContainer" containerID="300e129cdcf8b8d7284ead7f99b6e1b5e097938101e0073f735b5f97f1d6e063" Feb 22 00:22:31 crc kubenswrapper[4857]: I0222 00:22:31.950068 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8vwcq" Feb 22 00:22:31 crc kubenswrapper[4857]: I0222 00:22:31.980440 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8vwcq"] Feb 22 00:22:31 crc kubenswrapper[4857]: I0222 00:22:31.985972 4857 scope.go:117] "RemoveContainer" containerID="762157d0d8af1c189d65e10c5c3fdfbff510ca0be9532852d55617eefcf35ac0" Feb 22 00:22:31 crc kubenswrapper[4857]: I0222 00:22:31.986030 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8vwcq"] Feb 22 00:22:32 crc kubenswrapper[4857]: I0222 00:22:32.010678 4857 scope.go:117] "RemoveContainer" containerID="1160695d13a934e09ea55f880b0a0dafbf9abb0e63c78ca8248d931e89187aee" Feb 22 00:22:32 crc kubenswrapper[4857]: I0222 00:22:32.030075 4857 scope.go:117] "RemoveContainer" containerID="300e129cdcf8b8d7284ead7f99b6e1b5e097938101e0073f735b5f97f1d6e063" Feb 22 00:22:32 crc kubenswrapper[4857]: E0222 00:22:32.030521 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"300e129cdcf8b8d7284ead7f99b6e1b5e097938101e0073f735b5f97f1d6e063\": container with ID starting with 300e129cdcf8b8d7284ead7f99b6e1b5e097938101e0073f735b5f97f1d6e063 not found: ID does not exist" containerID="300e129cdcf8b8d7284ead7f99b6e1b5e097938101e0073f735b5f97f1d6e063" Feb 22 00:22:32 crc kubenswrapper[4857]: I0222 00:22:32.030553 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"300e129cdcf8b8d7284ead7f99b6e1b5e097938101e0073f735b5f97f1d6e063"} err="failed to get container status \"300e129cdcf8b8d7284ead7f99b6e1b5e097938101e0073f735b5f97f1d6e063\": rpc error: code = NotFound desc = could not find container \"300e129cdcf8b8d7284ead7f99b6e1b5e097938101e0073f735b5f97f1d6e063\": container with ID starting with 300e129cdcf8b8d7284ead7f99b6e1b5e097938101e0073f735b5f97f1d6e063 not found: ID does not exist" Feb 22 00:22:32 crc kubenswrapper[4857]: I0222 00:22:32.030574 4857 scope.go:117] "RemoveContainer" containerID="762157d0d8af1c189d65e10c5c3fdfbff510ca0be9532852d55617eefcf35ac0" Feb 22 00:22:32 crc kubenswrapper[4857]: E0222 00:22:32.031120 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"762157d0d8af1c189d65e10c5c3fdfbff510ca0be9532852d55617eefcf35ac0\": container with ID starting with 762157d0d8af1c189d65e10c5c3fdfbff510ca0be9532852d55617eefcf35ac0 not found: ID does not exist" containerID="762157d0d8af1c189d65e10c5c3fdfbff510ca0be9532852d55617eefcf35ac0" Feb 22 00:22:32 crc kubenswrapper[4857]: I0222 00:22:32.031178 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"762157d0d8af1c189d65e10c5c3fdfbff510ca0be9532852d55617eefcf35ac0"} err="failed to get container status \"762157d0d8af1c189d65e10c5c3fdfbff510ca0be9532852d55617eefcf35ac0\": rpc error: code = NotFound desc = could not find container \"762157d0d8af1c189d65e10c5c3fdfbff510ca0be9532852d55617eefcf35ac0\": container with ID starting with 762157d0d8af1c189d65e10c5c3fdfbff510ca0be9532852d55617eefcf35ac0 not found: ID does not exist" Feb 22 00:22:32 crc kubenswrapper[4857]: I0222 00:22:32.031191 4857 scope.go:117] "RemoveContainer" containerID="1160695d13a934e09ea55f880b0a0dafbf9abb0e63c78ca8248d931e89187aee" Feb 22 00:22:32 crc kubenswrapper[4857]: E0222 00:22:32.031505 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1160695d13a934e09ea55f880b0a0dafbf9abb0e63c78ca8248d931e89187aee\": container with ID starting with 1160695d13a934e09ea55f880b0a0dafbf9abb0e63c78ca8248d931e89187aee not found: ID does not exist" containerID="1160695d13a934e09ea55f880b0a0dafbf9abb0e63c78ca8248d931e89187aee" Feb 22 00:22:32 crc kubenswrapper[4857]: I0222 00:22:32.031529 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1160695d13a934e09ea55f880b0a0dafbf9abb0e63c78ca8248d931e89187aee"} err="failed to get container status \"1160695d13a934e09ea55f880b0a0dafbf9abb0e63c78ca8248d931e89187aee\": rpc error: code = NotFound desc = could not find container \"1160695d13a934e09ea55f880b0a0dafbf9abb0e63c78ca8248d931e89187aee\": container with ID starting with 1160695d13a934e09ea55f880b0a0dafbf9abb0e63c78ca8248d931e89187aee not found: ID does not exist" Feb 22 00:22:33 crc kubenswrapper[4857]: I0222 00:22:33.085819 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8bcd34c-0ae9-41cf-b579-83717bbe623e" path="/var/lib/kubelet/pods/f8bcd34c-0ae9-41cf-b579-83717bbe623e/volumes" Feb 22 00:22:38 crc kubenswrapper[4857]: E0222 00:22:38.078399 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:22:52 crc kubenswrapper[4857]: E0222 00:22:52.079324 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:23:06 crc kubenswrapper[4857]: E0222 00:23:06.095132 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:23:17 crc kubenswrapper[4857]: E0222 00:23:17.079608 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:23:31 crc kubenswrapper[4857]: E0222 00:23:31.079203 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:23:44 crc kubenswrapper[4857]: E0222 00:23:44.080664 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:23:54 crc kubenswrapper[4857]: I0222 00:23:54.312117 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-kqm64"] Feb 22 00:23:54 crc kubenswrapper[4857]: E0222 00:23:54.312758 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8bcd34c-0ae9-41cf-b579-83717bbe623e" containerName="extract-utilities" Feb 22 00:23:54 crc kubenswrapper[4857]: I0222 00:23:54.312773 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8bcd34c-0ae9-41cf-b579-83717bbe623e" containerName="extract-utilities" Feb 22 00:23:54 crc kubenswrapper[4857]: E0222 00:23:54.312792 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8bcd34c-0ae9-41cf-b579-83717bbe623e" containerName="extract-content" Feb 22 00:23:54 crc kubenswrapper[4857]: I0222 00:23:54.312800 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8bcd34c-0ae9-41cf-b579-83717bbe623e" containerName="extract-content" Feb 22 00:23:54 crc kubenswrapper[4857]: E0222 00:23:54.312820 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8bcd34c-0ae9-41cf-b579-83717bbe623e" containerName="registry-server" Feb 22 00:23:54 crc kubenswrapper[4857]: I0222 00:23:54.312829 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8bcd34c-0ae9-41cf-b579-83717bbe623e" containerName="registry-server" Feb 22 00:23:54 crc kubenswrapper[4857]: I0222 00:23:54.312968 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8bcd34c-0ae9-41cf-b579-83717bbe623e" containerName="registry-server" Feb 22 00:23:54 crc kubenswrapper[4857]: I0222 00:23:54.313492 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-kqm64" Feb 22 00:23:54 crc kubenswrapper[4857]: I0222 00:23:54.317609 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-kqm64"] Feb 22 00:23:54 crc kubenswrapper[4857]: I0222 00:23:54.369591 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qqd2\" (UniqueName: \"kubernetes.io/projected/d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76-kube-api-access-8qqd2\") pod \"infrawatch-operators-kqm64\" (UID: \"d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76\") " pod="service-telemetry/infrawatch-operators-kqm64" Feb 22 00:23:54 crc kubenswrapper[4857]: I0222 00:23:54.470813 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qqd2\" (UniqueName: \"kubernetes.io/projected/d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76-kube-api-access-8qqd2\") pod \"infrawatch-operators-kqm64\" (UID: \"d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76\") " pod="service-telemetry/infrawatch-operators-kqm64" Feb 22 00:23:54 crc kubenswrapper[4857]: I0222 00:23:54.496965 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qqd2\" (UniqueName: \"kubernetes.io/projected/d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76-kube-api-access-8qqd2\") pod \"infrawatch-operators-kqm64\" (UID: \"d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76\") " pod="service-telemetry/infrawatch-operators-kqm64" Feb 22 00:23:54 crc kubenswrapper[4857]: I0222 00:23:54.633551 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-kqm64" Feb 22 00:23:54 crc kubenswrapper[4857]: I0222 00:23:54.891328 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-kqm64"] Feb 22 00:23:54 crc kubenswrapper[4857]: E0222 00:23:54.934839 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest" Feb 22 00:23:54 crc kubenswrapper[4857]: E0222 00:23:54.935020 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8qqd2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infrawatch-operators-kqm64_service-telemetry(d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76): ErrImagePull: initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" logger="UnhandledError" Feb 22 00:23:54 crc kubenswrapper[4857]: E0222 00:23:54.936255 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:23:55 crc kubenswrapper[4857]: I0222 00:23:55.476046 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-kqm64" event={"ID":"d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76","Type":"ContainerStarted","Data":"b78c7793131419059375d2c332277c50668afa784860664110bb6bf819e41668"} Feb 22 00:23:55 crc kubenswrapper[4857]: E0222 00:23:55.478300 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:23:56 crc kubenswrapper[4857]: E0222 00:23:56.486548 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:23:58 crc kubenswrapper[4857]: E0222 00:23:58.079376 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:24:10 crc kubenswrapper[4857]: E0222 00:24:10.079628 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:24:11 crc kubenswrapper[4857]: E0222 00:24:11.113423 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest" Feb 22 00:24:11 crc kubenswrapper[4857]: E0222 00:24:11.113666 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8qqd2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infrawatch-operators-kqm64_service-telemetry(d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76): ErrImagePull: initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" logger="UnhandledError" Feb 22 00:24:11 crc kubenswrapper[4857]: E0222 00:24:11.114932 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:24:21 crc kubenswrapper[4857]: E0222 00:24:21.081261 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:24:25 crc kubenswrapper[4857]: E0222 00:24:25.081364 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:24:35 crc kubenswrapper[4857]: E0222 00:24:35.137174 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest" Feb 22 00:24:35 crc kubenswrapper[4857]: E0222 00:24:35.137933 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bqrvk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infrawatch-operators-9lk9c_service-telemetry(57fe7048-f646-407a-a625-0be29af288ff): ErrImagePull: initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" logger="UnhandledError" Feb 22 00:24:35 crc kubenswrapper[4857]: E0222 00:24:35.139160 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:24:37 crc kubenswrapper[4857]: E0222 00:24:37.101450 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest" Feb 22 00:24:37 crc kubenswrapper[4857]: E0222 00:24:37.103409 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8qqd2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infrawatch-operators-kqm64_service-telemetry(d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76): ErrImagePull: initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" logger="UnhandledError" Feb 22 00:24:37 crc kubenswrapper[4857]: E0222 00:24:37.104742 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:24:40 crc kubenswrapper[4857]: I0222 00:24:40.428496 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:24:40 crc kubenswrapper[4857]: I0222 00:24:40.428946 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:24:48 crc kubenswrapper[4857]: E0222 00:24:48.081207 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:24:48 crc kubenswrapper[4857]: E0222 00:24:48.081105 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:25:03 crc kubenswrapper[4857]: E0222 00:25:03.084804 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:25:03 crc kubenswrapper[4857]: E0222 00:25:03.084814 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:25:10 crc kubenswrapper[4857]: I0222 00:25:10.428131 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:25:10 crc kubenswrapper[4857]: I0222 00:25:10.428781 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:25:17 crc kubenswrapper[4857]: E0222 00:25:17.079621 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:25:18 crc kubenswrapper[4857]: E0222 00:25:18.079981 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:25:28 crc kubenswrapper[4857]: E0222 00:25:28.114590 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest" Feb 22 00:25:28 crc kubenswrapper[4857]: E0222 00:25:28.115388 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8qqd2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infrawatch-operators-kqm64_service-telemetry(d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76): ErrImagePull: initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" logger="UnhandledError" Feb 22 00:25:28 crc kubenswrapper[4857]: E0222 00:25:28.116693 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:25:30 crc kubenswrapper[4857]: E0222 00:25:30.079176 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:25:40 crc kubenswrapper[4857]: I0222 00:25:40.429091 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:25:40 crc kubenswrapper[4857]: I0222 00:25:40.430126 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:25:40 crc kubenswrapper[4857]: I0222 00:25:40.430196 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:25:40 crc kubenswrapper[4857]: I0222 00:25:40.431102 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a2bc37e253a36b33e8ab447c1e8c471a24067bf7db0e01c4c28c06ec0eef8f2c"} pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 22 00:25:40 crc kubenswrapper[4857]: I0222 00:25:40.431179 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" containerID="cri-o://a2bc37e253a36b33e8ab447c1e8c471a24067bf7db0e01c4c28c06ec0eef8f2c" gracePeriod=600 Feb 22 00:25:41 crc kubenswrapper[4857]: I0222 00:25:41.206135 4857 generic.go:334] "Generic (PLEG): container finished" podID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerID="a2bc37e253a36b33e8ab447c1e8c471a24067bf7db0e01c4c28c06ec0eef8f2c" exitCode=0 Feb 22 00:25:41 crc kubenswrapper[4857]: I0222 00:25:41.206227 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" event={"ID":"2ed2fde8-1449-491b-ae21-2fe11ab008ae","Type":"ContainerDied","Data":"a2bc37e253a36b33e8ab447c1e8c471a24067bf7db0e01c4c28c06ec0eef8f2c"} Feb 22 00:25:41 crc kubenswrapper[4857]: I0222 00:25:41.206561 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" event={"ID":"2ed2fde8-1449-491b-ae21-2fe11ab008ae","Type":"ContainerStarted","Data":"c2027575d139fdf13c928a69109637e8223bf66d3edbc5bc76086f1d56269d6e"} Feb 22 00:25:41 crc kubenswrapper[4857]: I0222 00:25:41.206592 4857 scope.go:117] "RemoveContainer" containerID="d7b4b58f072c91cd783115813cd320a04ba05440e4f3d243be01bcab4bd1e785" Feb 22 00:25:42 crc kubenswrapper[4857]: E0222 00:25:42.078813 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:25:43 crc kubenswrapper[4857]: E0222 00:25:43.081335 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:25:57 crc kubenswrapper[4857]: E0222 00:25:57.081335 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:25:57 crc kubenswrapper[4857]: E0222 00:25:57.081358 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:26:08 crc kubenswrapper[4857]: E0222 00:26:08.080173 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:26:09 crc kubenswrapper[4857]: E0222 00:26:09.084911 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:26:21 crc kubenswrapper[4857]: E0222 00:26:21.086634 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:26:23 crc kubenswrapper[4857]: E0222 00:26:23.083822 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:26:33 crc kubenswrapper[4857]: E0222 00:26:33.086751 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:26:34 crc kubenswrapper[4857]: E0222 00:26:34.079830 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:26:46 crc kubenswrapper[4857]: E0222 00:26:46.080305 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:26:48 crc kubenswrapper[4857]: E0222 00:26:48.079693 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:26:59 crc kubenswrapper[4857]: I0222 00:26:59.081098 4857 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 22 00:26:59 crc kubenswrapper[4857]: E0222 00:26:59.127718 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest" Feb 22 00:26:59 crc kubenswrapper[4857]: E0222 00:26:59.128239 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8qqd2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infrawatch-operators-kqm64_service-telemetry(d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76): ErrImagePull: initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" logger="UnhandledError" Feb 22 00:26:59 crc kubenswrapper[4857]: E0222 00:26:59.129490 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:27:01 crc kubenswrapper[4857]: E0222 00:27:01.079126 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:27:14 crc kubenswrapper[4857]: E0222 00:27:14.081059 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:27:16 crc kubenswrapper[4857]: E0222 00:27:16.079785 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:27:25 crc kubenswrapper[4857]: E0222 00:27:25.080945 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:27:30 crc kubenswrapper[4857]: E0222 00:27:30.081916 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:27:37 crc kubenswrapper[4857]: E0222 00:27:37.080064 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:27:40 crc kubenswrapper[4857]: I0222 00:27:40.428523 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:27:40 crc kubenswrapper[4857]: I0222 00:27:40.428824 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:27:45 crc kubenswrapper[4857]: E0222 00:27:45.080465 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:27:50 crc kubenswrapper[4857]: E0222 00:27:50.080562 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:28:00 crc kubenswrapper[4857]: E0222 00:28:00.079810 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:28:03 crc kubenswrapper[4857]: E0222 00:28:03.083372 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:28:10 crc kubenswrapper[4857]: I0222 00:28:10.428840 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:28:10 crc kubenswrapper[4857]: I0222 00:28:10.429736 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:28:14 crc kubenswrapper[4857]: E0222 00:28:14.079135 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:28:17 crc kubenswrapper[4857]: E0222 00:28:17.078969 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:28:29 crc kubenswrapper[4857]: E0222 00:28:29.079580 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:28:31 crc kubenswrapper[4857]: E0222 00:28:31.083376 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:28:40 crc kubenswrapper[4857]: I0222 00:28:40.428451 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:28:40 crc kubenswrapper[4857]: I0222 00:28:40.428513 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:28:40 crc kubenswrapper[4857]: I0222 00:28:40.428557 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:28:40 crc kubenswrapper[4857]: I0222 00:28:40.429103 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c2027575d139fdf13c928a69109637e8223bf66d3edbc5bc76086f1d56269d6e"} pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 22 00:28:40 crc kubenswrapper[4857]: I0222 00:28:40.429150 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" containerID="cri-o://c2027575d139fdf13c928a69109637e8223bf66d3edbc5bc76086f1d56269d6e" gracePeriod=600 Feb 22 00:28:41 crc kubenswrapper[4857]: I0222 00:28:41.454879 4857 generic.go:334] "Generic (PLEG): container finished" podID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerID="c2027575d139fdf13c928a69109637e8223bf66d3edbc5bc76086f1d56269d6e" exitCode=0 Feb 22 00:28:41 crc kubenswrapper[4857]: I0222 00:28:41.455102 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" event={"ID":"2ed2fde8-1449-491b-ae21-2fe11ab008ae","Type":"ContainerDied","Data":"c2027575d139fdf13c928a69109637e8223bf66d3edbc5bc76086f1d56269d6e"} Feb 22 00:28:41 crc kubenswrapper[4857]: I0222 00:28:41.455436 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" event={"ID":"2ed2fde8-1449-491b-ae21-2fe11ab008ae","Type":"ContainerStarted","Data":"1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6"} Feb 22 00:28:41 crc kubenswrapper[4857]: I0222 00:28:41.455458 4857 scope.go:117] "RemoveContainer" containerID="a2bc37e253a36b33e8ab447c1e8c471a24067bf7db0e01c4c28c06ec0eef8f2c" Feb 22 00:28:43 crc kubenswrapper[4857]: E0222 00:28:43.083131 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:28:46 crc kubenswrapper[4857]: E0222 00:28:46.078866 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:28:55 crc kubenswrapper[4857]: E0222 00:28:55.078941 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:28:58 crc kubenswrapper[4857]: E0222 00:28:58.081016 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:29:09 crc kubenswrapper[4857]: E0222 00:29:09.080005 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:29:10 crc kubenswrapper[4857]: E0222 00:29:10.079087 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:29:21 crc kubenswrapper[4857]: E0222 00:29:21.079909 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:29:24 crc kubenswrapper[4857]: E0222 00:29:24.079573 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:29:35 crc kubenswrapper[4857]: E0222 00:29:35.097131 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:29:36 crc kubenswrapper[4857]: E0222 00:29:36.079209 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:29:48 crc kubenswrapper[4857]: E0222 00:29:48.134384 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest" Feb 22 00:29:48 crc kubenswrapper[4857]: E0222 00:29:48.135375 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8qqd2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infrawatch-operators-kqm64_service-telemetry(d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76): ErrImagePull: initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" logger="UnhandledError" Feb 22 00:29:48 crc kubenswrapper[4857]: E0222 00:29:48.136628 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:29:49 crc kubenswrapper[4857]: E0222 00:29:49.107253 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest" Feb 22 00:29:49 crc kubenswrapper[4857]: E0222 00:29:49.107439 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bqrvk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infrawatch-operators-9lk9c_service-telemetry(57fe7048-f646-407a-a625-0be29af288ff): ErrImagePull: initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" logger="UnhandledError" Feb 22 00:29:49 crc kubenswrapper[4857]: E0222 00:29:49.108894 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:29:51 crc kubenswrapper[4857]: I0222 00:29:51.124271 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-m5q8l/must-gather-hrvpz"] Feb 22 00:29:51 crc kubenswrapper[4857]: I0222 00:29:51.125390 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5q8l/must-gather-hrvpz" Feb 22 00:29:51 crc kubenswrapper[4857]: I0222 00:29:51.127586 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-m5q8l"/"openshift-service-ca.crt" Feb 22 00:29:51 crc kubenswrapper[4857]: I0222 00:29:51.127939 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-m5q8l"/"kube-root-ca.crt" Feb 22 00:29:51 crc kubenswrapper[4857]: I0222 00:29:51.192790 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-m5q8l/must-gather-hrvpz"] Feb 22 00:29:51 crc kubenswrapper[4857]: I0222 00:29:51.238895 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e9d3b041-2e01-4f32-a808-aa2917afa4cc-must-gather-output\") pod \"must-gather-hrvpz\" (UID: \"e9d3b041-2e01-4f32-a808-aa2917afa4cc\") " pod="openshift-must-gather-m5q8l/must-gather-hrvpz" Feb 22 00:29:51 crc kubenswrapper[4857]: I0222 00:29:51.239854 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-449kd\" (UniqueName: \"kubernetes.io/projected/e9d3b041-2e01-4f32-a808-aa2917afa4cc-kube-api-access-449kd\") pod \"must-gather-hrvpz\" (UID: \"e9d3b041-2e01-4f32-a808-aa2917afa4cc\") " pod="openshift-must-gather-m5q8l/must-gather-hrvpz" Feb 22 00:29:51 crc kubenswrapper[4857]: I0222 00:29:51.340416 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e9d3b041-2e01-4f32-a808-aa2917afa4cc-must-gather-output\") pod \"must-gather-hrvpz\" (UID: \"e9d3b041-2e01-4f32-a808-aa2917afa4cc\") " pod="openshift-must-gather-m5q8l/must-gather-hrvpz" Feb 22 00:29:51 crc kubenswrapper[4857]: I0222 00:29:51.340524 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-449kd\" (UniqueName: \"kubernetes.io/projected/e9d3b041-2e01-4f32-a808-aa2917afa4cc-kube-api-access-449kd\") pod \"must-gather-hrvpz\" (UID: \"e9d3b041-2e01-4f32-a808-aa2917afa4cc\") " pod="openshift-must-gather-m5q8l/must-gather-hrvpz" Feb 22 00:29:51 crc kubenswrapper[4857]: I0222 00:29:51.340860 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e9d3b041-2e01-4f32-a808-aa2917afa4cc-must-gather-output\") pod \"must-gather-hrvpz\" (UID: \"e9d3b041-2e01-4f32-a808-aa2917afa4cc\") " pod="openshift-must-gather-m5q8l/must-gather-hrvpz" Feb 22 00:29:51 crc kubenswrapper[4857]: I0222 00:29:51.359512 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-449kd\" (UniqueName: \"kubernetes.io/projected/e9d3b041-2e01-4f32-a808-aa2917afa4cc-kube-api-access-449kd\") pod \"must-gather-hrvpz\" (UID: \"e9d3b041-2e01-4f32-a808-aa2917afa4cc\") " pod="openshift-must-gather-m5q8l/must-gather-hrvpz" Feb 22 00:29:51 crc kubenswrapper[4857]: I0222 00:29:51.441360 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5q8l/must-gather-hrvpz" Feb 22 00:29:51 crc kubenswrapper[4857]: I0222 00:29:51.871438 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-m5q8l/must-gather-hrvpz"] Feb 22 00:29:51 crc kubenswrapper[4857]: W0222 00:29:51.874642 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9d3b041_2e01_4f32_a808_aa2917afa4cc.slice/crio-3a3a6c27ab5dbc5dd8e5137384bbf94dac3e5d4be483ae24550c7358c9c77ebe WatchSource:0}: Error finding container 3a3a6c27ab5dbc5dd8e5137384bbf94dac3e5d4be483ae24550c7358c9c77ebe: Status 404 returned error can't find the container with id 3a3a6c27ab5dbc5dd8e5137384bbf94dac3e5d4be483ae24550c7358c9c77ebe Feb 22 00:29:51 crc kubenswrapper[4857]: I0222 00:29:51.934975 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m5q8l/must-gather-hrvpz" event={"ID":"e9d3b041-2e01-4f32-a808-aa2917afa4cc","Type":"ContainerStarted","Data":"3a3a6c27ab5dbc5dd8e5137384bbf94dac3e5d4be483ae24550c7358c9c77ebe"} Feb 22 00:29:57 crc kubenswrapper[4857]: I0222 00:29:57.966468 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m5q8l/must-gather-hrvpz" event={"ID":"e9d3b041-2e01-4f32-a808-aa2917afa4cc","Type":"ContainerStarted","Data":"f09c0ee065abfaac10a34149d5fc2b9e8aef4cee5111e34e656ae241fb316d5d"} Feb 22 00:29:57 crc kubenswrapper[4857]: I0222 00:29:57.966995 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m5q8l/must-gather-hrvpz" event={"ID":"e9d3b041-2e01-4f32-a808-aa2917afa4cc","Type":"ContainerStarted","Data":"8d91a7b601bf98db6f63be2f2093a86da356307d7f246a206b8c6c88fa69d6d3"} Feb 22 00:29:57 crc kubenswrapper[4857]: I0222 00:29:57.981048 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-m5q8l/must-gather-hrvpz" podStartSLOduration=1.514382598 podStartE2EDuration="6.981010571s" podCreationTimestamp="2026-02-22 00:29:51 +0000 UTC" firstStartedPulling="2026-02-22 00:29:51.877164036 +0000 UTC m=+1399.515893289" lastFinishedPulling="2026-02-22 00:29:57.343792009 +0000 UTC m=+1404.982521262" observedRunningTime="2026-02-22 00:29:57.978546025 +0000 UTC m=+1405.617275278" watchObservedRunningTime="2026-02-22 00:29:57.981010571 +0000 UTC m=+1405.619739824" Feb 22 00:30:00 crc kubenswrapper[4857]: I0222 00:30:00.138283 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm"] Feb 22 00:30:00 crc kubenswrapper[4857]: I0222 00:30:00.139575 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm" Feb 22 00:30:00 crc kubenswrapper[4857]: I0222 00:30:00.142942 4857 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 22 00:30:00 crc kubenswrapper[4857]: I0222 00:30:00.143144 4857 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 22 00:30:00 crc kubenswrapper[4857]: I0222 00:30:00.149180 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm"] Feb 22 00:30:00 crc kubenswrapper[4857]: I0222 00:30:00.217482 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4tzr\" (UniqueName: \"kubernetes.io/projected/9dd1df69-d878-4876-a92e-be2ae1e313b9-kube-api-access-h4tzr\") pod \"collect-profiles-29528670-scmnm\" (UID: \"9dd1df69-d878-4876-a92e-be2ae1e313b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm" Feb 22 00:30:00 crc kubenswrapper[4857]: I0222 00:30:00.217624 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9dd1df69-d878-4876-a92e-be2ae1e313b9-config-volume\") pod \"collect-profiles-29528670-scmnm\" (UID: \"9dd1df69-d878-4876-a92e-be2ae1e313b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm" Feb 22 00:30:00 crc kubenswrapper[4857]: I0222 00:30:00.217796 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9dd1df69-d878-4876-a92e-be2ae1e313b9-secret-volume\") pod \"collect-profiles-29528670-scmnm\" (UID: \"9dd1df69-d878-4876-a92e-be2ae1e313b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm" Feb 22 00:30:00 crc kubenswrapper[4857]: I0222 00:30:00.318972 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4tzr\" (UniqueName: \"kubernetes.io/projected/9dd1df69-d878-4876-a92e-be2ae1e313b9-kube-api-access-h4tzr\") pod \"collect-profiles-29528670-scmnm\" (UID: \"9dd1df69-d878-4876-a92e-be2ae1e313b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm" Feb 22 00:30:00 crc kubenswrapper[4857]: I0222 00:30:00.319042 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9dd1df69-d878-4876-a92e-be2ae1e313b9-config-volume\") pod \"collect-profiles-29528670-scmnm\" (UID: \"9dd1df69-d878-4876-a92e-be2ae1e313b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm" Feb 22 00:30:00 crc kubenswrapper[4857]: I0222 00:30:00.319125 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9dd1df69-d878-4876-a92e-be2ae1e313b9-secret-volume\") pod \"collect-profiles-29528670-scmnm\" (UID: \"9dd1df69-d878-4876-a92e-be2ae1e313b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm" Feb 22 00:30:00 crc kubenswrapper[4857]: I0222 00:30:00.320698 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9dd1df69-d878-4876-a92e-be2ae1e313b9-config-volume\") pod \"collect-profiles-29528670-scmnm\" (UID: \"9dd1df69-d878-4876-a92e-be2ae1e313b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm" Feb 22 00:30:00 crc kubenswrapper[4857]: I0222 00:30:00.325888 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9dd1df69-d878-4876-a92e-be2ae1e313b9-secret-volume\") pod \"collect-profiles-29528670-scmnm\" (UID: \"9dd1df69-d878-4876-a92e-be2ae1e313b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm" Feb 22 00:30:00 crc kubenswrapper[4857]: I0222 00:30:00.342931 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4tzr\" (UniqueName: \"kubernetes.io/projected/9dd1df69-d878-4876-a92e-be2ae1e313b9-kube-api-access-h4tzr\") pod \"collect-profiles-29528670-scmnm\" (UID: \"9dd1df69-d878-4876-a92e-be2ae1e313b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm" Feb 22 00:30:00 crc kubenswrapper[4857]: I0222 00:30:00.460298 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm" Feb 22 00:30:00 crc kubenswrapper[4857]: I0222 00:30:00.661858 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm"] Feb 22 00:30:00 crc kubenswrapper[4857]: I0222 00:30:00.984301 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm" event={"ID":"9dd1df69-d878-4876-a92e-be2ae1e313b9","Type":"ContainerStarted","Data":"861299ee7fca689c4484f322d8c69e754cb2d05102a93f3f3e168e84ed78dd0f"} Feb 22 00:30:00 crc kubenswrapper[4857]: I0222 00:30:00.984650 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm" event={"ID":"9dd1df69-d878-4876-a92e-be2ae1e313b9","Type":"ContainerStarted","Data":"113d732b8b228429ca4a5edaaeea6e5f8d16341b538a5d1acabe064c69a02fb7"} Feb 22 00:30:01 crc kubenswrapper[4857]: I0222 00:30:01.003993 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm" podStartSLOduration=1.003976976 podStartE2EDuration="1.003976976s" podCreationTimestamp="2026-02-22 00:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-22 00:30:00.99898816 +0000 UTC m=+1408.637717413" watchObservedRunningTime="2026-02-22 00:30:01.003976976 +0000 UTC m=+1408.642706229" Feb 22 00:30:01 crc kubenswrapper[4857]: E0222 00:30:01.078588 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:30:02 crc kubenswrapper[4857]: I0222 00:30:02.000301 4857 generic.go:334] "Generic (PLEG): container finished" podID="9dd1df69-d878-4876-a92e-be2ae1e313b9" containerID="861299ee7fca689c4484f322d8c69e754cb2d05102a93f3f3e168e84ed78dd0f" exitCode=0 Feb 22 00:30:02 crc kubenswrapper[4857]: I0222 00:30:02.000477 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm" event={"ID":"9dd1df69-d878-4876-a92e-be2ae1e313b9","Type":"ContainerDied","Data":"861299ee7fca689c4484f322d8c69e754cb2d05102a93f3f3e168e84ed78dd0f"} Feb 22 00:30:03 crc kubenswrapper[4857]: E0222 00:30:03.081752 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:30:03 crc kubenswrapper[4857]: I0222 00:30:03.268000 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm" Feb 22 00:30:03 crc kubenswrapper[4857]: I0222 00:30:03.453947 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9dd1df69-d878-4876-a92e-be2ae1e313b9-secret-volume\") pod \"9dd1df69-d878-4876-a92e-be2ae1e313b9\" (UID: \"9dd1df69-d878-4876-a92e-be2ae1e313b9\") " Feb 22 00:30:03 crc kubenswrapper[4857]: I0222 00:30:03.454017 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4tzr\" (UniqueName: \"kubernetes.io/projected/9dd1df69-d878-4876-a92e-be2ae1e313b9-kube-api-access-h4tzr\") pod \"9dd1df69-d878-4876-a92e-be2ae1e313b9\" (UID: \"9dd1df69-d878-4876-a92e-be2ae1e313b9\") " Feb 22 00:30:03 crc kubenswrapper[4857]: I0222 00:30:03.454110 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9dd1df69-d878-4876-a92e-be2ae1e313b9-config-volume\") pod \"9dd1df69-d878-4876-a92e-be2ae1e313b9\" (UID: \"9dd1df69-d878-4876-a92e-be2ae1e313b9\") " Feb 22 00:30:03 crc kubenswrapper[4857]: I0222 00:30:03.454970 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dd1df69-d878-4876-a92e-be2ae1e313b9-config-volume" (OuterVolumeSpecName: "config-volume") pod "9dd1df69-d878-4876-a92e-be2ae1e313b9" (UID: "9dd1df69-d878-4876-a92e-be2ae1e313b9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 22 00:30:03 crc kubenswrapper[4857]: I0222 00:30:03.459203 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dd1df69-d878-4876-a92e-be2ae1e313b9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9dd1df69-d878-4876-a92e-be2ae1e313b9" (UID: "9dd1df69-d878-4876-a92e-be2ae1e313b9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 22 00:30:03 crc kubenswrapper[4857]: I0222 00:30:03.465364 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dd1df69-d878-4876-a92e-be2ae1e313b9-kube-api-access-h4tzr" (OuterVolumeSpecName: "kube-api-access-h4tzr") pod "9dd1df69-d878-4876-a92e-be2ae1e313b9" (UID: "9dd1df69-d878-4876-a92e-be2ae1e313b9"). InnerVolumeSpecName "kube-api-access-h4tzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:30:03 crc kubenswrapper[4857]: I0222 00:30:03.555750 4857 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9dd1df69-d878-4876-a92e-be2ae1e313b9-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 22 00:30:03 crc kubenswrapper[4857]: I0222 00:30:03.556044 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4tzr\" (UniqueName: \"kubernetes.io/projected/9dd1df69-d878-4876-a92e-be2ae1e313b9-kube-api-access-h4tzr\") on node \"crc\" DevicePath \"\"" Feb 22 00:30:03 crc kubenswrapper[4857]: I0222 00:30:03.556168 4857 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9dd1df69-d878-4876-a92e-be2ae1e313b9-config-volume\") on node \"crc\" DevicePath \"\"" Feb 22 00:30:04 crc kubenswrapper[4857]: I0222 00:30:04.013520 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm" event={"ID":"9dd1df69-d878-4876-a92e-be2ae1e313b9","Type":"ContainerDied","Data":"113d732b8b228429ca4a5edaaeea6e5f8d16341b538a5d1acabe064c69a02fb7"} Feb 22 00:30:04 crc kubenswrapper[4857]: I0222 00:30:04.013572 4857 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="113d732b8b228429ca4a5edaaeea6e5f8d16341b538a5d1acabe064c69a02fb7" Feb 22 00:30:04 crc kubenswrapper[4857]: I0222 00:30:04.013870 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29528670-scmnm" Feb 22 00:30:14 crc kubenswrapper[4857]: E0222 00:30:14.079321 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:30:15 crc kubenswrapper[4857]: E0222 00:30:15.079221 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:30:27 crc kubenswrapper[4857]: E0222 00:30:27.078918 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:30:29 crc kubenswrapper[4857]: E0222 00:30:29.079014 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:30:35 crc kubenswrapper[4857]: I0222 00:30:35.978863 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-4drbd_e6ba7763-9bd5-45dd-a9a9-d47dc7f381ba/control-plane-machine-set-operator/0.log" Feb 22 00:30:36 crc kubenswrapper[4857]: I0222 00:30:36.088088 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-cwlf7_f9bce83c-ae40-482f-88d2-7de36a1a7ea5/kube-rbac-proxy/0.log" Feb 22 00:30:36 crc kubenswrapper[4857]: I0222 00:30:36.125970 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-cwlf7_f9bce83c-ae40-482f-88d2-7de36a1a7ea5/machine-api-operator/0.log" Feb 22 00:30:39 crc kubenswrapper[4857]: E0222 00:30:39.079743 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:30:40 crc kubenswrapper[4857]: I0222 00:30:40.428912 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:30:40 crc kubenswrapper[4857]: I0222 00:30:40.428968 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:30:43 crc kubenswrapper[4857]: E0222 00:30:43.082199 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:30:46 crc kubenswrapper[4857]: I0222 00:30:46.291060 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-545d4d4674-6nr72_63f9c6da-3d8d-4ae7-b619-52db9af8a21f/cert-manager-controller/0.log" Feb 22 00:30:46 crc kubenswrapper[4857]: I0222 00:30:46.531860 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-5545bd876-gpl5d_d68808f6-f5f0-452f-9036-abce9664d3e7/cert-manager-cainjector/0.log" Feb 22 00:30:46 crc kubenswrapper[4857]: I0222 00:30:46.592109 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-6888856db4-qphbj_f9b6b93d-c3b5-4165-a9a6-28274a241b7a/cert-manager-webhook/0.log" Feb 22 00:30:51 crc kubenswrapper[4857]: E0222 00:30:51.080485 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:30:54 crc kubenswrapper[4857]: E0222 00:30:54.078645 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:31:00 crc kubenswrapper[4857]: I0222 00:31:00.930021 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-6zdkm_056b8239-b8e8-4610-b38d-8e4cd84ddda5/prometheus-operator/0.log" Feb 22 00:31:01 crc kubenswrapper[4857]: I0222 00:31:01.075705 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx_faf26ac1-5a62-47cd-bf57-fe48ffb196cc/prometheus-operator-admission-webhook/0.log" Feb 22 00:31:01 crc kubenswrapper[4857]: I0222 00:31:01.105767 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c_65eb5aaa-f178-4680-b66b-a468cd347d3a/prometheus-operator-admission-webhook/0.log" Feb 22 00:31:01 crc kubenswrapper[4857]: I0222 00:31:01.220339 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-kcx7r_bae29f5b-7bae-41c4-a5a0-b5f6e9f57729/operator/0.log" Feb 22 00:31:01 crc kubenswrapper[4857]: I0222 00:31:01.277363 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-w8x45_177b7675-71d3-4c71-8781-4e352edc0044/perses-operator/0.log" Feb 22 00:31:02 crc kubenswrapper[4857]: E0222 00:31:02.079439 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:31:08 crc kubenswrapper[4857]: E0222 00:31:08.078999 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:31:10 crc kubenswrapper[4857]: I0222 00:31:10.428426 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:31:10 crc kubenswrapper[4857]: I0222 00:31:10.428769 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:31:13 crc kubenswrapper[4857]: E0222 00:31:13.081904 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:31:13 crc kubenswrapper[4857]: I0222 00:31:13.295020 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6_15ca2571-3005-4eac-80ce-71db200efbfa/util/0.log" Feb 22 00:31:13 crc kubenswrapper[4857]: I0222 00:31:13.457707 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6_15ca2571-3005-4eac-80ce-71db200efbfa/pull/0.log" Feb 22 00:31:13 crc kubenswrapper[4857]: I0222 00:31:13.477119 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6_15ca2571-3005-4eac-80ce-71db200efbfa/util/0.log" Feb 22 00:31:13 crc kubenswrapper[4857]: I0222 00:31:13.495475 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6_15ca2571-3005-4eac-80ce-71db200efbfa/pull/0.log" Feb 22 00:31:13 crc kubenswrapper[4857]: I0222 00:31:13.659800 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6_15ca2571-3005-4eac-80ce-71db200efbfa/util/0.log" Feb 22 00:31:13 crc kubenswrapper[4857]: I0222 00:31:13.663632 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6_15ca2571-3005-4eac-80ce-71db200efbfa/extract/0.log" Feb 22 00:31:13 crc kubenswrapper[4857]: I0222 00:31:13.669005 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f16x4j6_15ca2571-3005-4eac-80ce-71db200efbfa/pull/0.log" Feb 22 00:31:13 crc kubenswrapper[4857]: I0222 00:31:13.807320 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8_8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc/util/0.log" Feb 22 00:31:13 crc kubenswrapper[4857]: I0222 00:31:13.946966 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8_8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc/util/0.log" Feb 22 00:31:13 crc kubenswrapper[4857]: I0222 00:31:13.980427 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8_8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc/pull/0.log" Feb 22 00:31:14 crc kubenswrapper[4857]: I0222 00:31:14.012167 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8_8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc/pull/0.log" Feb 22 00:31:14 crc kubenswrapper[4857]: I0222 00:31:14.129127 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8_8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc/util/0.log" Feb 22 00:31:14 crc kubenswrapper[4857]: I0222 00:31:14.131799 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8_8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc/extract/0.log" Feb 22 00:31:14 crc kubenswrapper[4857]: I0222 00:31:14.158053 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e58l9m8_8f8c4b9d-d2d6-4c5a-8dae-48602b19f9fc/pull/0.log" Feb 22 00:31:14 crc kubenswrapper[4857]: I0222 00:31:14.287414 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q_489bd4f1-d8e2-4db7-a4b8-76755e099093/util/0.log" Feb 22 00:31:14 crc kubenswrapper[4857]: I0222 00:31:14.451982 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q_489bd4f1-d8e2-4db7-a4b8-76755e099093/util/0.log" Feb 22 00:31:14 crc kubenswrapper[4857]: I0222 00:31:14.452971 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q_489bd4f1-d8e2-4db7-a4b8-76755e099093/pull/0.log" Feb 22 00:31:14 crc kubenswrapper[4857]: I0222 00:31:14.461224 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q_489bd4f1-d8e2-4db7-a4b8-76755e099093/pull/0.log" Feb 22 00:31:14 crc kubenswrapper[4857]: I0222 00:31:14.588187 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q_489bd4f1-d8e2-4db7-a4b8-76755e099093/pull/0.log" Feb 22 00:31:14 crc kubenswrapper[4857]: I0222 00:31:14.625481 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q_489bd4f1-d8e2-4db7-a4b8-76755e099093/util/0.log" Feb 22 00:31:14 crc kubenswrapper[4857]: I0222 00:31:14.673612 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08znk7q_489bd4f1-d8e2-4db7-a4b8-76755e099093/extract/0.log" Feb 22 00:31:14 crc kubenswrapper[4857]: I0222 00:31:14.754974 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nx245_361dcb13-b990-403e-814b-cf0d11668bbf/extract-utilities/0.log" Feb 22 00:31:14 crc kubenswrapper[4857]: I0222 00:31:14.884679 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nx245_361dcb13-b990-403e-814b-cf0d11668bbf/extract-utilities/0.log" Feb 22 00:31:14 crc kubenswrapper[4857]: I0222 00:31:14.910181 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nx245_361dcb13-b990-403e-814b-cf0d11668bbf/extract-content/0.log" Feb 22 00:31:14 crc kubenswrapper[4857]: I0222 00:31:14.911134 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nx245_361dcb13-b990-403e-814b-cf0d11668bbf/extract-content/0.log" Feb 22 00:31:15 crc kubenswrapper[4857]: I0222 00:31:15.032418 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nx245_361dcb13-b990-403e-814b-cf0d11668bbf/extract-content/0.log" Feb 22 00:31:15 crc kubenswrapper[4857]: I0222 00:31:15.053045 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nx245_361dcb13-b990-403e-814b-cf0d11668bbf/extract-utilities/0.log" Feb 22 00:31:15 crc kubenswrapper[4857]: I0222 00:31:15.244733 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fwkxt_dc489b84-7090-4294-84f5-77edb82e6ca0/extract-utilities/0.log" Feb 22 00:31:15 crc kubenswrapper[4857]: I0222 00:31:15.330086 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-nx245_361dcb13-b990-403e-814b-cf0d11668bbf/registry-server/0.log" Feb 22 00:31:15 crc kubenswrapper[4857]: I0222 00:31:15.399959 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fwkxt_dc489b84-7090-4294-84f5-77edb82e6ca0/extract-utilities/0.log" Feb 22 00:31:15 crc kubenswrapper[4857]: I0222 00:31:15.432132 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fwkxt_dc489b84-7090-4294-84f5-77edb82e6ca0/extract-content/0.log" Feb 22 00:31:15 crc kubenswrapper[4857]: I0222 00:31:15.437462 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fwkxt_dc489b84-7090-4294-84f5-77edb82e6ca0/extract-content/0.log" Feb 22 00:31:15 crc kubenswrapper[4857]: I0222 00:31:15.625962 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fwkxt_dc489b84-7090-4294-84f5-77edb82e6ca0/extract-utilities/0.log" Feb 22 00:31:15 crc kubenswrapper[4857]: I0222 00:31:15.633317 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fwkxt_dc489b84-7090-4294-84f5-77edb82e6ca0/extract-content/0.log" Feb 22 00:31:15 crc kubenswrapper[4857]: I0222 00:31:15.781237 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fwkxt_dc489b84-7090-4294-84f5-77edb82e6ca0/registry-server/0.log" Feb 22 00:31:15 crc kubenswrapper[4857]: I0222 00:31:15.815492 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4cnjs_aec64a2e-ff83-4d35-a6cc-691c906ee0d9/marketplace-operator/3.log" Feb 22 00:31:15 crc kubenswrapper[4857]: I0222 00:31:15.854790 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4cnjs_aec64a2e-ff83-4d35-a6cc-691c906ee0d9/marketplace-operator/2.log" Feb 22 00:31:15 crc kubenswrapper[4857]: I0222 00:31:15.945783 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xrdpk_ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2/extract-utilities/0.log" Feb 22 00:31:16 crc kubenswrapper[4857]: I0222 00:31:16.096732 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xrdpk_ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2/extract-content/0.log" Feb 22 00:31:16 crc kubenswrapper[4857]: I0222 00:31:16.103889 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xrdpk_ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2/extract-utilities/0.log" Feb 22 00:31:16 crc kubenswrapper[4857]: I0222 00:31:16.108682 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xrdpk_ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2/extract-content/0.log" Feb 22 00:31:16 crc kubenswrapper[4857]: I0222 00:31:16.228289 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xrdpk_ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2/extract-utilities/0.log" Feb 22 00:31:16 crc kubenswrapper[4857]: I0222 00:31:16.253407 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xrdpk_ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2/extract-content/0.log" Feb 22 00:31:16 crc kubenswrapper[4857]: I0222 00:31:16.440851 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-xrdpk_ca1ea5d7-d3fa-4aa9-ad7e-393011f9d5f2/registry-server/0.log" Feb 22 00:31:19 crc kubenswrapper[4857]: E0222 00:31:19.078952 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:31:27 crc kubenswrapper[4857]: I0222 00:31:27.016507 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-68d8d8f674-88kwx_faf26ac1-5a62-47cd-bf57-fe48ffb196cc/prometheus-operator-admission-webhook/0.log" Feb 22 00:31:27 crc kubenswrapper[4857]: I0222 00:31:27.045126 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-68d8d8f674-t9k2c_65eb5aaa-f178-4680-b66b-a468cd347d3a/prometheus-operator-admission-webhook/0.log" Feb 22 00:31:27 crc kubenswrapper[4857]: I0222 00:31:27.049381 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-6zdkm_056b8239-b8e8-4610-b38d-8e4cd84ddda5/prometheus-operator/0.log" Feb 22 00:31:27 crc kubenswrapper[4857]: E0222 00:31:27.078820 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:31:27 crc kubenswrapper[4857]: I0222 00:31:27.156301 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-kcx7r_bae29f5b-7bae-41c4-a5a0-b5f6e9f57729/operator/0.log" Feb 22 00:31:27 crc kubenswrapper[4857]: I0222 00:31:27.197899 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-w8x45_177b7675-71d3-4c71-8781-4e352edc0044/perses-operator/0.log" Feb 22 00:31:31 crc kubenswrapper[4857]: E0222 00:31:31.078708 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:31:40 crc kubenswrapper[4857]: I0222 00:31:40.428518 4857 patch_prober.go:28] interesting pod/machine-config-daemon-gw6k5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 22 00:31:40 crc kubenswrapper[4857]: I0222 00:31:40.429732 4857 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 22 00:31:40 crc kubenswrapper[4857]: I0222 00:31:40.429925 4857 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" Feb 22 00:31:40 crc kubenswrapper[4857]: I0222 00:31:40.430471 4857 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6"} pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 22 00:31:40 crc kubenswrapper[4857]: I0222 00:31:40.430594 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerName="machine-config-daemon" containerID="cri-o://1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" gracePeriod=600 Feb 22 00:31:40 crc kubenswrapper[4857]: E0222 00:31:40.567613 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:31:41 crc kubenswrapper[4857]: I0222 00:31:41.524827 4857 generic.go:334] "Generic (PLEG): container finished" podID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" exitCode=0 Feb 22 00:31:41 crc kubenswrapper[4857]: I0222 00:31:41.525560 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" event={"ID":"2ed2fde8-1449-491b-ae21-2fe11ab008ae","Type":"ContainerDied","Data":"1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6"} Feb 22 00:31:41 crc kubenswrapper[4857]: I0222 00:31:41.525735 4857 scope.go:117] "RemoveContainer" containerID="c2027575d139fdf13c928a69109637e8223bf66d3edbc5bc76086f1d56269d6e" Feb 22 00:31:41 crc kubenswrapper[4857]: I0222 00:31:41.526332 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:31:41 crc kubenswrapper[4857]: E0222 00:31:41.526672 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:31:42 crc kubenswrapper[4857]: E0222 00:31:42.079780 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:31:42 crc kubenswrapper[4857]: E0222 00:31:42.080076 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:31:54 crc kubenswrapper[4857]: I0222 00:31:54.077428 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:31:54 crc kubenswrapper[4857]: E0222 00:31:54.078423 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:31:54 crc kubenswrapper[4857]: E0222 00:31:54.079651 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:31:56 crc kubenswrapper[4857]: E0222 00:31:56.079490 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:32:08 crc kubenswrapper[4857]: I0222 00:32:08.077838 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:32:08 crc kubenswrapper[4857]: E0222 00:32:08.079003 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:32:08 crc kubenswrapper[4857]: E0222 00:32:08.089515 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:32:09 crc kubenswrapper[4857]: E0222 00:32:09.082200 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:32:17 crc kubenswrapper[4857]: I0222 00:32:17.806108 4857 generic.go:334] "Generic (PLEG): container finished" podID="e9d3b041-2e01-4f32-a808-aa2917afa4cc" containerID="8d91a7b601bf98db6f63be2f2093a86da356307d7f246a206b8c6c88fa69d6d3" exitCode=0 Feb 22 00:32:17 crc kubenswrapper[4857]: I0222 00:32:17.806161 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m5q8l/must-gather-hrvpz" event={"ID":"e9d3b041-2e01-4f32-a808-aa2917afa4cc","Type":"ContainerDied","Data":"8d91a7b601bf98db6f63be2f2093a86da356307d7f246a206b8c6c88fa69d6d3"} Feb 22 00:32:17 crc kubenswrapper[4857]: I0222 00:32:17.807083 4857 scope.go:117] "RemoveContainer" containerID="8d91a7b601bf98db6f63be2f2093a86da356307d7f246a206b8c6c88fa69d6d3" Feb 22 00:32:18 crc kubenswrapper[4857]: I0222 00:32:18.383319 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m5q8l_must-gather-hrvpz_e9d3b041-2e01-4f32-a808-aa2917afa4cc/gather/0.log" Feb 22 00:32:20 crc kubenswrapper[4857]: I0222 00:32:20.078214 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:32:20 crc kubenswrapper[4857]: E0222 00:32:20.079060 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:32:20 crc kubenswrapper[4857]: E0222 00:32:20.079972 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:32:22 crc kubenswrapper[4857]: E0222 00:32:22.078264 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.245727 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-m5q8l/must-gather-hrvpz"] Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.246996 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-m5q8l/must-gather-hrvpz" podUID="e9d3b041-2e01-4f32-a808-aa2917afa4cc" containerName="copy" containerID="cri-o://f09c0ee065abfaac10a34149d5fc2b9e8aef4cee5111e34e656ae241fb316d5d" gracePeriod=2 Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.252165 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-m5q8l/must-gather-hrvpz"] Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.589362 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m5q8l_must-gather-hrvpz_e9d3b041-2e01-4f32-a808-aa2917afa4cc/copy/0.log" Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.590269 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5q8l/must-gather-hrvpz" Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.616341 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e9d3b041-2e01-4f32-a808-aa2917afa4cc-must-gather-output\") pod \"e9d3b041-2e01-4f32-a808-aa2917afa4cc\" (UID: \"e9d3b041-2e01-4f32-a808-aa2917afa4cc\") " Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.616494 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-449kd\" (UniqueName: \"kubernetes.io/projected/e9d3b041-2e01-4f32-a808-aa2917afa4cc-kube-api-access-449kd\") pod \"e9d3b041-2e01-4f32-a808-aa2917afa4cc\" (UID: \"e9d3b041-2e01-4f32-a808-aa2917afa4cc\") " Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.624852 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9d3b041-2e01-4f32-a808-aa2917afa4cc-kube-api-access-449kd" (OuterVolumeSpecName: "kube-api-access-449kd") pod "e9d3b041-2e01-4f32-a808-aa2917afa4cc" (UID: "e9d3b041-2e01-4f32-a808-aa2917afa4cc"). InnerVolumeSpecName "kube-api-access-449kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.676114 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9d3b041-2e01-4f32-a808-aa2917afa4cc-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e9d3b041-2e01-4f32-a808-aa2917afa4cc" (UID: "e9d3b041-2e01-4f32-a808-aa2917afa4cc"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.718320 4857 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e9d3b041-2e01-4f32-a808-aa2917afa4cc-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.718360 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-449kd\" (UniqueName: \"kubernetes.io/projected/e9d3b041-2e01-4f32-a808-aa2917afa4cc-kube-api-access-449kd\") on node \"crc\" DevicePath \"\"" Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.867350 4857 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m5q8l_must-gather-hrvpz_e9d3b041-2e01-4f32-a808-aa2917afa4cc/copy/0.log" Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.867968 4857 generic.go:334] "Generic (PLEG): container finished" podID="e9d3b041-2e01-4f32-a808-aa2917afa4cc" containerID="f09c0ee065abfaac10a34149d5fc2b9e8aef4cee5111e34e656ae241fb316d5d" exitCode=143 Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.868020 4857 scope.go:117] "RemoveContainer" containerID="f09c0ee065abfaac10a34149d5fc2b9e8aef4cee5111e34e656ae241fb316d5d" Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.868090 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m5q8l/must-gather-hrvpz" Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.890101 4857 scope.go:117] "RemoveContainer" containerID="8d91a7b601bf98db6f63be2f2093a86da356307d7f246a206b8c6c88fa69d6d3" Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.934400 4857 scope.go:117] "RemoveContainer" containerID="f09c0ee065abfaac10a34149d5fc2b9e8aef4cee5111e34e656ae241fb316d5d" Feb 22 00:32:25 crc kubenswrapper[4857]: E0222 00:32:25.934900 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f09c0ee065abfaac10a34149d5fc2b9e8aef4cee5111e34e656ae241fb316d5d\": container with ID starting with f09c0ee065abfaac10a34149d5fc2b9e8aef4cee5111e34e656ae241fb316d5d not found: ID does not exist" containerID="f09c0ee065abfaac10a34149d5fc2b9e8aef4cee5111e34e656ae241fb316d5d" Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.934942 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f09c0ee065abfaac10a34149d5fc2b9e8aef4cee5111e34e656ae241fb316d5d"} err="failed to get container status \"f09c0ee065abfaac10a34149d5fc2b9e8aef4cee5111e34e656ae241fb316d5d\": rpc error: code = NotFound desc = could not find container \"f09c0ee065abfaac10a34149d5fc2b9e8aef4cee5111e34e656ae241fb316d5d\": container with ID starting with f09c0ee065abfaac10a34149d5fc2b9e8aef4cee5111e34e656ae241fb316d5d not found: ID does not exist" Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.934972 4857 scope.go:117] "RemoveContainer" containerID="8d91a7b601bf98db6f63be2f2093a86da356307d7f246a206b8c6c88fa69d6d3" Feb 22 00:32:25 crc kubenswrapper[4857]: E0222 00:32:25.935384 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d91a7b601bf98db6f63be2f2093a86da356307d7f246a206b8c6c88fa69d6d3\": container with ID starting with 8d91a7b601bf98db6f63be2f2093a86da356307d7f246a206b8c6c88fa69d6d3 not found: ID does not exist" containerID="8d91a7b601bf98db6f63be2f2093a86da356307d7f246a206b8c6c88fa69d6d3" Feb 22 00:32:25 crc kubenswrapper[4857]: I0222 00:32:25.935426 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d91a7b601bf98db6f63be2f2093a86da356307d7f246a206b8c6c88fa69d6d3"} err="failed to get container status \"8d91a7b601bf98db6f63be2f2093a86da356307d7f246a206b8c6c88fa69d6d3\": rpc error: code = NotFound desc = could not find container \"8d91a7b601bf98db6f63be2f2093a86da356307d7f246a206b8c6c88fa69d6d3\": container with ID starting with 8d91a7b601bf98db6f63be2f2093a86da356307d7f246a206b8c6c88fa69d6d3 not found: ID does not exist" Feb 22 00:32:27 crc kubenswrapper[4857]: I0222 00:32:27.088959 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9d3b041-2e01-4f32-a808-aa2917afa4cc" path="/var/lib/kubelet/pods/e9d3b041-2e01-4f32-a808-aa2917afa4cc/volumes" Feb 22 00:32:31 crc kubenswrapper[4857]: I0222 00:32:31.076965 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:32:31 crc kubenswrapper[4857]: E0222 00:32:31.077844 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:32:32 crc kubenswrapper[4857]: E0222 00:32:32.080611 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:32:32 crc kubenswrapper[4857]: I0222 00:32:32.675641 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-db425"] Feb 22 00:32:32 crc kubenswrapper[4857]: E0222 00:32:32.676795 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dd1df69-d878-4876-a92e-be2ae1e313b9" containerName="collect-profiles" Feb 22 00:32:32 crc kubenswrapper[4857]: I0222 00:32:32.676817 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dd1df69-d878-4876-a92e-be2ae1e313b9" containerName="collect-profiles" Feb 22 00:32:32 crc kubenswrapper[4857]: E0222 00:32:32.676835 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9d3b041-2e01-4f32-a808-aa2917afa4cc" containerName="gather" Feb 22 00:32:32 crc kubenswrapper[4857]: I0222 00:32:32.676841 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9d3b041-2e01-4f32-a808-aa2917afa4cc" containerName="gather" Feb 22 00:32:32 crc kubenswrapper[4857]: E0222 00:32:32.676854 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9d3b041-2e01-4f32-a808-aa2917afa4cc" containerName="copy" Feb 22 00:32:32 crc kubenswrapper[4857]: I0222 00:32:32.676861 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9d3b041-2e01-4f32-a808-aa2917afa4cc" containerName="copy" Feb 22 00:32:32 crc kubenswrapper[4857]: I0222 00:32:32.676951 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dd1df69-d878-4876-a92e-be2ae1e313b9" containerName="collect-profiles" Feb 22 00:32:32 crc kubenswrapper[4857]: I0222 00:32:32.676965 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9d3b041-2e01-4f32-a808-aa2917afa4cc" containerName="gather" Feb 22 00:32:32 crc kubenswrapper[4857]: I0222 00:32:32.676975 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9d3b041-2e01-4f32-a808-aa2917afa4cc" containerName="copy" Feb 22 00:32:32 crc kubenswrapper[4857]: I0222 00:32:32.677910 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-db425" Feb 22 00:32:32 crc kubenswrapper[4857]: I0222 00:32:32.688371 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-db425"] Feb 22 00:32:32 crc kubenswrapper[4857]: I0222 00:32:32.773577 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94f39804-e8b7-45af-8160-523bc874c3f1-catalog-content\") pod \"certified-operators-db425\" (UID: \"94f39804-e8b7-45af-8160-523bc874c3f1\") " pod="openshift-marketplace/certified-operators-db425" Feb 22 00:32:32 crc kubenswrapper[4857]: I0222 00:32:32.773654 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94f39804-e8b7-45af-8160-523bc874c3f1-utilities\") pod \"certified-operators-db425\" (UID: \"94f39804-e8b7-45af-8160-523bc874c3f1\") " pod="openshift-marketplace/certified-operators-db425" Feb 22 00:32:32 crc kubenswrapper[4857]: I0222 00:32:32.773737 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66hjq\" (UniqueName: \"kubernetes.io/projected/94f39804-e8b7-45af-8160-523bc874c3f1-kube-api-access-66hjq\") pod \"certified-operators-db425\" (UID: \"94f39804-e8b7-45af-8160-523bc874c3f1\") " pod="openshift-marketplace/certified-operators-db425" Feb 22 00:32:32 crc kubenswrapper[4857]: I0222 00:32:32.874703 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66hjq\" (UniqueName: \"kubernetes.io/projected/94f39804-e8b7-45af-8160-523bc874c3f1-kube-api-access-66hjq\") pod \"certified-operators-db425\" (UID: \"94f39804-e8b7-45af-8160-523bc874c3f1\") " pod="openshift-marketplace/certified-operators-db425" Feb 22 00:32:32 crc kubenswrapper[4857]: I0222 00:32:32.874796 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94f39804-e8b7-45af-8160-523bc874c3f1-catalog-content\") pod \"certified-operators-db425\" (UID: \"94f39804-e8b7-45af-8160-523bc874c3f1\") " pod="openshift-marketplace/certified-operators-db425" Feb 22 00:32:32 crc kubenswrapper[4857]: I0222 00:32:32.874847 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94f39804-e8b7-45af-8160-523bc874c3f1-utilities\") pod \"certified-operators-db425\" (UID: \"94f39804-e8b7-45af-8160-523bc874c3f1\") " pod="openshift-marketplace/certified-operators-db425" Feb 22 00:32:32 crc kubenswrapper[4857]: I0222 00:32:32.875371 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94f39804-e8b7-45af-8160-523bc874c3f1-utilities\") pod \"certified-operators-db425\" (UID: \"94f39804-e8b7-45af-8160-523bc874c3f1\") " pod="openshift-marketplace/certified-operators-db425" Feb 22 00:32:32 crc kubenswrapper[4857]: I0222 00:32:32.875473 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94f39804-e8b7-45af-8160-523bc874c3f1-catalog-content\") pod \"certified-operators-db425\" (UID: \"94f39804-e8b7-45af-8160-523bc874c3f1\") " pod="openshift-marketplace/certified-operators-db425" Feb 22 00:32:32 crc kubenswrapper[4857]: I0222 00:32:32.892942 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66hjq\" (UniqueName: \"kubernetes.io/projected/94f39804-e8b7-45af-8160-523bc874c3f1-kube-api-access-66hjq\") pod \"certified-operators-db425\" (UID: \"94f39804-e8b7-45af-8160-523bc874c3f1\") " pod="openshift-marketplace/certified-operators-db425" Feb 22 00:32:33 crc kubenswrapper[4857]: I0222 00:32:33.006972 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-db425" Feb 22 00:32:33 crc kubenswrapper[4857]: I0222 00:32:33.239989 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-db425"] Feb 22 00:32:33 crc kubenswrapper[4857]: W0222 00:32:33.240225 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94f39804_e8b7_45af_8160_523bc874c3f1.slice/crio-aa3d33feee52a4d04730be9a4a008dc30d080cb8953d953ed8deac4d6a4b6776 WatchSource:0}: Error finding container aa3d33feee52a4d04730be9a4a008dc30d080cb8953d953ed8deac4d6a4b6776: Status 404 returned error can't find the container with id aa3d33feee52a4d04730be9a4a008dc30d080cb8953d953ed8deac4d6a4b6776 Feb 22 00:32:33 crc kubenswrapper[4857]: I0222 00:32:33.923208 4857 generic.go:334] "Generic (PLEG): container finished" podID="94f39804-e8b7-45af-8160-523bc874c3f1" containerID="81a4bfc6fa7eea6fd2705e30d35883fd926ea524f965c88636ebc1d2d43fdd92" exitCode=0 Feb 22 00:32:33 crc kubenswrapper[4857]: I0222 00:32:33.923276 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-db425" event={"ID":"94f39804-e8b7-45af-8160-523bc874c3f1","Type":"ContainerDied","Data":"81a4bfc6fa7eea6fd2705e30d35883fd926ea524f965c88636ebc1d2d43fdd92"} Feb 22 00:32:33 crc kubenswrapper[4857]: I0222 00:32:33.923509 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-db425" event={"ID":"94f39804-e8b7-45af-8160-523bc874c3f1","Type":"ContainerStarted","Data":"aa3d33feee52a4d04730be9a4a008dc30d080cb8953d953ed8deac4d6a4b6776"} Feb 22 00:32:33 crc kubenswrapper[4857]: I0222 00:32:33.932992 4857 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 22 00:32:34 crc kubenswrapper[4857]: I0222 00:32:34.930063 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-db425" event={"ID":"94f39804-e8b7-45af-8160-523bc874c3f1","Type":"ContainerStarted","Data":"5759e7c23db72cf48d60d94a24a725caf5535cd7c38d6a9c373cd095b96b5889"} Feb 22 00:32:35 crc kubenswrapper[4857]: E0222 00:32:35.079497 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:32:35 crc kubenswrapper[4857]: I0222 00:32:35.936476 4857 generic.go:334] "Generic (PLEG): container finished" podID="94f39804-e8b7-45af-8160-523bc874c3f1" containerID="5759e7c23db72cf48d60d94a24a725caf5535cd7c38d6a9c373cd095b96b5889" exitCode=0 Feb 22 00:32:35 crc kubenswrapper[4857]: I0222 00:32:35.936554 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-db425" event={"ID":"94f39804-e8b7-45af-8160-523bc874c3f1","Type":"ContainerDied","Data":"5759e7c23db72cf48d60d94a24a725caf5535cd7c38d6a9c373cd095b96b5889"} Feb 22 00:32:36 crc kubenswrapper[4857]: I0222 00:32:36.945428 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-db425" event={"ID":"94f39804-e8b7-45af-8160-523bc874c3f1","Type":"ContainerStarted","Data":"9886b47907f44d0f1099fe9e23f9692af4b0000a68a57715763c672a17bdb2f9"} Feb 22 00:32:36 crc kubenswrapper[4857]: I0222 00:32:36.966961 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-db425" podStartSLOduration=2.571782947 podStartE2EDuration="4.966942749s" podCreationTimestamp="2026-02-22 00:32:32 +0000 UTC" firstStartedPulling="2026-02-22 00:32:33.932678285 +0000 UTC m=+1561.571407538" lastFinishedPulling="2026-02-22 00:32:36.327838077 +0000 UTC m=+1563.966567340" observedRunningTime="2026-02-22 00:32:36.960857259 +0000 UTC m=+1564.599586522" watchObservedRunningTime="2026-02-22 00:32:36.966942749 +0000 UTC m=+1564.605672012" Feb 22 00:32:42 crc kubenswrapper[4857]: I0222 00:32:42.078059 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:32:42 crc kubenswrapper[4857]: E0222 00:32:42.078732 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:32:43 crc kubenswrapper[4857]: I0222 00:32:43.007549 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-db425" Feb 22 00:32:43 crc kubenswrapper[4857]: I0222 00:32:43.007616 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-db425" Feb 22 00:32:43 crc kubenswrapper[4857]: I0222 00:32:43.048404 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-db425" Feb 22 00:32:44 crc kubenswrapper[4857]: I0222 00:32:44.021305 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-db425" Feb 22 00:32:44 crc kubenswrapper[4857]: I0222 00:32:44.072120 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-db425"] Feb 22 00:32:45 crc kubenswrapper[4857]: I0222 00:32:45.996784 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-db425" podUID="94f39804-e8b7-45af-8160-523bc874c3f1" containerName="registry-server" containerID="cri-o://9886b47907f44d0f1099fe9e23f9692af4b0000a68a57715763c672a17bdb2f9" gracePeriod=2 Feb 22 00:32:46 crc kubenswrapper[4857]: E0222 00:32:46.078689 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:32:47 crc kubenswrapper[4857]: I0222 00:32:47.005334 4857 generic.go:334] "Generic (PLEG): container finished" podID="94f39804-e8b7-45af-8160-523bc874c3f1" containerID="9886b47907f44d0f1099fe9e23f9692af4b0000a68a57715763c672a17bdb2f9" exitCode=0 Feb 22 00:32:47 crc kubenswrapper[4857]: I0222 00:32:47.005387 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-db425" event={"ID":"94f39804-e8b7-45af-8160-523bc874c3f1","Type":"ContainerDied","Data":"9886b47907f44d0f1099fe9e23f9692af4b0000a68a57715763c672a17bdb2f9"} Feb 22 00:32:47 crc kubenswrapper[4857]: E0222 00:32:47.078592 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:32:47 crc kubenswrapper[4857]: I0222 00:32:47.510210 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-db425" Feb 22 00:32:47 crc kubenswrapper[4857]: I0222 00:32:47.661833 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66hjq\" (UniqueName: \"kubernetes.io/projected/94f39804-e8b7-45af-8160-523bc874c3f1-kube-api-access-66hjq\") pod \"94f39804-e8b7-45af-8160-523bc874c3f1\" (UID: \"94f39804-e8b7-45af-8160-523bc874c3f1\") " Feb 22 00:32:47 crc kubenswrapper[4857]: I0222 00:32:47.661955 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94f39804-e8b7-45af-8160-523bc874c3f1-utilities\") pod \"94f39804-e8b7-45af-8160-523bc874c3f1\" (UID: \"94f39804-e8b7-45af-8160-523bc874c3f1\") " Feb 22 00:32:47 crc kubenswrapper[4857]: I0222 00:32:47.662015 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94f39804-e8b7-45af-8160-523bc874c3f1-catalog-content\") pod \"94f39804-e8b7-45af-8160-523bc874c3f1\" (UID: \"94f39804-e8b7-45af-8160-523bc874c3f1\") " Feb 22 00:32:47 crc kubenswrapper[4857]: I0222 00:32:47.663410 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94f39804-e8b7-45af-8160-523bc874c3f1-utilities" (OuterVolumeSpecName: "utilities") pod "94f39804-e8b7-45af-8160-523bc874c3f1" (UID: "94f39804-e8b7-45af-8160-523bc874c3f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:32:47 crc kubenswrapper[4857]: I0222 00:32:47.667401 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94f39804-e8b7-45af-8160-523bc874c3f1-kube-api-access-66hjq" (OuterVolumeSpecName: "kube-api-access-66hjq") pod "94f39804-e8b7-45af-8160-523bc874c3f1" (UID: "94f39804-e8b7-45af-8160-523bc874c3f1"). InnerVolumeSpecName "kube-api-access-66hjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:32:47 crc kubenswrapper[4857]: I0222 00:32:47.714947 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94f39804-e8b7-45af-8160-523bc874c3f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94f39804-e8b7-45af-8160-523bc874c3f1" (UID: "94f39804-e8b7-45af-8160-523bc874c3f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:32:47 crc kubenswrapper[4857]: I0222 00:32:47.764260 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94f39804-e8b7-45af-8160-523bc874c3f1-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:32:47 crc kubenswrapper[4857]: I0222 00:32:47.764310 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94f39804-e8b7-45af-8160-523bc874c3f1-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:32:47 crc kubenswrapper[4857]: I0222 00:32:47.764325 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66hjq\" (UniqueName: \"kubernetes.io/projected/94f39804-e8b7-45af-8160-523bc874c3f1-kube-api-access-66hjq\") on node \"crc\" DevicePath \"\"" Feb 22 00:32:48 crc kubenswrapper[4857]: I0222 00:32:48.014211 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-db425" event={"ID":"94f39804-e8b7-45af-8160-523bc874c3f1","Type":"ContainerDied","Data":"aa3d33feee52a4d04730be9a4a008dc30d080cb8953d953ed8deac4d6a4b6776"} Feb 22 00:32:48 crc kubenswrapper[4857]: I0222 00:32:48.014293 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-db425" Feb 22 00:32:48 crc kubenswrapper[4857]: I0222 00:32:48.014308 4857 scope.go:117] "RemoveContainer" containerID="9886b47907f44d0f1099fe9e23f9692af4b0000a68a57715763c672a17bdb2f9" Feb 22 00:32:48 crc kubenswrapper[4857]: I0222 00:32:48.030575 4857 scope.go:117] "RemoveContainer" containerID="5759e7c23db72cf48d60d94a24a725caf5535cd7c38d6a9c373cd095b96b5889" Feb 22 00:32:48 crc kubenswrapper[4857]: I0222 00:32:48.070936 4857 scope.go:117] "RemoveContainer" containerID="81a4bfc6fa7eea6fd2705e30d35883fd926ea524f965c88636ebc1d2d43fdd92" Feb 22 00:32:48 crc kubenswrapper[4857]: I0222 00:32:48.076178 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-db425"] Feb 22 00:32:48 crc kubenswrapper[4857]: I0222 00:32:48.091264 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-db425"] Feb 22 00:32:49 crc kubenswrapper[4857]: I0222 00:32:49.086122 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94f39804-e8b7-45af-8160-523bc874c3f1" path="/var/lib/kubelet/pods/94f39804-e8b7-45af-8160-523bc874c3f1/volumes" Feb 22 00:32:54 crc kubenswrapper[4857]: I0222 00:32:54.078007 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:32:54 crc kubenswrapper[4857]: E0222 00:32:54.078520 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:33:00 crc kubenswrapper[4857]: E0222 00:33:00.079662 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:33:01 crc kubenswrapper[4857]: E0222 00:33:01.081498 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:33:09 crc kubenswrapper[4857]: I0222 00:33:09.077172 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:33:09 crc kubenswrapper[4857]: E0222 00:33:09.077938 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:33:14 crc kubenswrapper[4857]: E0222 00:33:14.080194 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:33:15 crc kubenswrapper[4857]: E0222 00:33:15.080232 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:33:22 crc kubenswrapper[4857]: I0222 00:33:22.077843 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:33:22 crc kubenswrapper[4857]: E0222 00:33:22.078695 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:33:25 crc kubenswrapper[4857]: E0222 00:33:25.080733 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:33:28 crc kubenswrapper[4857]: E0222 00:33:28.079486 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:33:28 crc kubenswrapper[4857]: I0222 00:33:28.995802 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qqlfh"] Feb 22 00:33:28 crc kubenswrapper[4857]: E0222 00:33:28.996518 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94f39804-e8b7-45af-8160-523bc874c3f1" containerName="extract-content" Feb 22 00:33:28 crc kubenswrapper[4857]: I0222 00:33:28.996536 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="94f39804-e8b7-45af-8160-523bc874c3f1" containerName="extract-content" Feb 22 00:33:28 crc kubenswrapper[4857]: E0222 00:33:28.996556 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94f39804-e8b7-45af-8160-523bc874c3f1" containerName="registry-server" Feb 22 00:33:28 crc kubenswrapper[4857]: I0222 00:33:28.996565 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="94f39804-e8b7-45af-8160-523bc874c3f1" containerName="registry-server" Feb 22 00:33:28 crc kubenswrapper[4857]: E0222 00:33:28.996575 4857 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94f39804-e8b7-45af-8160-523bc874c3f1" containerName="extract-utilities" Feb 22 00:33:28 crc kubenswrapper[4857]: I0222 00:33:28.996584 4857 state_mem.go:107] "Deleted CPUSet assignment" podUID="94f39804-e8b7-45af-8160-523bc874c3f1" containerName="extract-utilities" Feb 22 00:33:28 crc kubenswrapper[4857]: I0222 00:33:28.996720 4857 memory_manager.go:354] "RemoveStaleState removing state" podUID="94f39804-e8b7-45af-8160-523bc874c3f1" containerName="registry-server" Feb 22 00:33:28 crc kubenswrapper[4857]: I0222 00:33:28.997782 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qqlfh" Feb 22 00:33:29 crc kubenswrapper[4857]: I0222 00:33:29.015206 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qqlfh"] Feb 22 00:33:29 crc kubenswrapper[4857]: I0222 00:33:29.153722 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz4lj\" (UniqueName: \"kubernetes.io/projected/7cc49fb4-2ca3-49dc-8a95-30998f97e95b-kube-api-access-tz4lj\") pod \"redhat-operators-qqlfh\" (UID: \"7cc49fb4-2ca3-49dc-8a95-30998f97e95b\") " pod="openshift-marketplace/redhat-operators-qqlfh" Feb 22 00:33:29 crc kubenswrapper[4857]: I0222 00:33:29.153829 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cc49fb4-2ca3-49dc-8a95-30998f97e95b-utilities\") pod \"redhat-operators-qqlfh\" (UID: \"7cc49fb4-2ca3-49dc-8a95-30998f97e95b\") " pod="openshift-marketplace/redhat-operators-qqlfh" Feb 22 00:33:29 crc kubenswrapper[4857]: I0222 00:33:29.153871 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cc49fb4-2ca3-49dc-8a95-30998f97e95b-catalog-content\") pod \"redhat-operators-qqlfh\" (UID: \"7cc49fb4-2ca3-49dc-8a95-30998f97e95b\") " pod="openshift-marketplace/redhat-operators-qqlfh" Feb 22 00:33:29 crc kubenswrapper[4857]: I0222 00:33:29.254788 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cc49fb4-2ca3-49dc-8a95-30998f97e95b-utilities\") pod \"redhat-operators-qqlfh\" (UID: \"7cc49fb4-2ca3-49dc-8a95-30998f97e95b\") " pod="openshift-marketplace/redhat-operators-qqlfh" Feb 22 00:33:29 crc kubenswrapper[4857]: I0222 00:33:29.254859 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cc49fb4-2ca3-49dc-8a95-30998f97e95b-catalog-content\") pod \"redhat-operators-qqlfh\" (UID: \"7cc49fb4-2ca3-49dc-8a95-30998f97e95b\") " pod="openshift-marketplace/redhat-operators-qqlfh" Feb 22 00:33:29 crc kubenswrapper[4857]: I0222 00:33:29.254917 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz4lj\" (UniqueName: \"kubernetes.io/projected/7cc49fb4-2ca3-49dc-8a95-30998f97e95b-kube-api-access-tz4lj\") pod \"redhat-operators-qqlfh\" (UID: \"7cc49fb4-2ca3-49dc-8a95-30998f97e95b\") " pod="openshift-marketplace/redhat-operators-qqlfh" Feb 22 00:33:29 crc kubenswrapper[4857]: I0222 00:33:29.255352 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cc49fb4-2ca3-49dc-8a95-30998f97e95b-utilities\") pod \"redhat-operators-qqlfh\" (UID: \"7cc49fb4-2ca3-49dc-8a95-30998f97e95b\") " pod="openshift-marketplace/redhat-operators-qqlfh" Feb 22 00:33:29 crc kubenswrapper[4857]: I0222 00:33:29.255370 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cc49fb4-2ca3-49dc-8a95-30998f97e95b-catalog-content\") pod \"redhat-operators-qqlfh\" (UID: \"7cc49fb4-2ca3-49dc-8a95-30998f97e95b\") " pod="openshift-marketplace/redhat-operators-qqlfh" Feb 22 00:33:29 crc kubenswrapper[4857]: I0222 00:33:29.274937 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz4lj\" (UniqueName: \"kubernetes.io/projected/7cc49fb4-2ca3-49dc-8a95-30998f97e95b-kube-api-access-tz4lj\") pod \"redhat-operators-qqlfh\" (UID: \"7cc49fb4-2ca3-49dc-8a95-30998f97e95b\") " pod="openshift-marketplace/redhat-operators-qqlfh" Feb 22 00:33:29 crc kubenswrapper[4857]: I0222 00:33:29.336016 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qqlfh" Feb 22 00:33:29 crc kubenswrapper[4857]: I0222 00:33:29.625111 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qqlfh"] Feb 22 00:33:30 crc kubenswrapper[4857]: I0222 00:33:30.268652 4857 generic.go:334] "Generic (PLEG): container finished" podID="7cc49fb4-2ca3-49dc-8a95-30998f97e95b" containerID="aaab1793ec017cde895a0ea311fb16c4a9ddb68484a853e2949effb7b20e9644" exitCode=0 Feb 22 00:33:30 crc kubenswrapper[4857]: I0222 00:33:30.268754 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqlfh" event={"ID":"7cc49fb4-2ca3-49dc-8a95-30998f97e95b","Type":"ContainerDied","Data":"aaab1793ec017cde895a0ea311fb16c4a9ddb68484a853e2949effb7b20e9644"} Feb 22 00:33:30 crc kubenswrapper[4857]: I0222 00:33:30.268931 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqlfh" event={"ID":"7cc49fb4-2ca3-49dc-8a95-30998f97e95b","Type":"ContainerStarted","Data":"52334db83e4b813c31cd32ad5767eccce3dd6399e56a7b47ce1d8116967ec518"} Feb 22 00:33:31 crc kubenswrapper[4857]: I0222 00:33:31.275242 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqlfh" event={"ID":"7cc49fb4-2ca3-49dc-8a95-30998f97e95b","Type":"ContainerStarted","Data":"54db3241cd351152e586e0a4934bc27134069108503f0ad2f6832d8daa165748"} Feb 22 00:33:31 crc kubenswrapper[4857]: I0222 00:33:31.989030 4857 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nsd69"] Feb 22 00:33:31 crc kubenswrapper[4857]: I0222 00:33:31.990446 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nsd69" Feb 22 00:33:32 crc kubenswrapper[4857]: I0222 00:33:32.026121 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nsd69"] Feb 22 00:33:32 crc kubenswrapper[4857]: I0222 00:33:32.089968 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc082b19-503f-4c5e-af76-99fe7e2bf0fd-catalog-content\") pod \"community-operators-nsd69\" (UID: \"fc082b19-503f-4c5e-af76-99fe7e2bf0fd\") " pod="openshift-marketplace/community-operators-nsd69" Feb 22 00:33:32 crc kubenswrapper[4857]: I0222 00:33:32.090054 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc082b19-503f-4c5e-af76-99fe7e2bf0fd-utilities\") pod \"community-operators-nsd69\" (UID: \"fc082b19-503f-4c5e-af76-99fe7e2bf0fd\") " pod="openshift-marketplace/community-operators-nsd69" Feb 22 00:33:32 crc kubenswrapper[4857]: I0222 00:33:32.090111 4857 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrmz7\" (UniqueName: \"kubernetes.io/projected/fc082b19-503f-4c5e-af76-99fe7e2bf0fd-kube-api-access-lrmz7\") pod \"community-operators-nsd69\" (UID: \"fc082b19-503f-4c5e-af76-99fe7e2bf0fd\") " pod="openshift-marketplace/community-operators-nsd69" Feb 22 00:33:32 crc kubenswrapper[4857]: I0222 00:33:32.190852 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc082b19-503f-4c5e-af76-99fe7e2bf0fd-catalog-content\") pod \"community-operators-nsd69\" (UID: \"fc082b19-503f-4c5e-af76-99fe7e2bf0fd\") " pod="openshift-marketplace/community-operators-nsd69" Feb 22 00:33:32 crc kubenswrapper[4857]: I0222 00:33:32.190932 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc082b19-503f-4c5e-af76-99fe7e2bf0fd-utilities\") pod \"community-operators-nsd69\" (UID: \"fc082b19-503f-4c5e-af76-99fe7e2bf0fd\") " pod="openshift-marketplace/community-operators-nsd69" Feb 22 00:33:32 crc kubenswrapper[4857]: I0222 00:33:32.190980 4857 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrmz7\" (UniqueName: \"kubernetes.io/projected/fc082b19-503f-4c5e-af76-99fe7e2bf0fd-kube-api-access-lrmz7\") pod \"community-operators-nsd69\" (UID: \"fc082b19-503f-4c5e-af76-99fe7e2bf0fd\") " pod="openshift-marketplace/community-operators-nsd69" Feb 22 00:33:32 crc kubenswrapper[4857]: I0222 00:33:32.191414 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc082b19-503f-4c5e-af76-99fe7e2bf0fd-catalog-content\") pod \"community-operators-nsd69\" (UID: \"fc082b19-503f-4c5e-af76-99fe7e2bf0fd\") " pod="openshift-marketplace/community-operators-nsd69" Feb 22 00:33:32 crc kubenswrapper[4857]: I0222 00:33:32.191458 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc082b19-503f-4c5e-af76-99fe7e2bf0fd-utilities\") pod \"community-operators-nsd69\" (UID: \"fc082b19-503f-4c5e-af76-99fe7e2bf0fd\") " pod="openshift-marketplace/community-operators-nsd69" Feb 22 00:33:32 crc kubenswrapper[4857]: I0222 00:33:32.211945 4857 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrmz7\" (UniqueName: \"kubernetes.io/projected/fc082b19-503f-4c5e-af76-99fe7e2bf0fd-kube-api-access-lrmz7\") pod \"community-operators-nsd69\" (UID: \"fc082b19-503f-4c5e-af76-99fe7e2bf0fd\") " pod="openshift-marketplace/community-operators-nsd69" Feb 22 00:33:32 crc kubenswrapper[4857]: I0222 00:33:32.282423 4857 generic.go:334] "Generic (PLEG): container finished" podID="7cc49fb4-2ca3-49dc-8a95-30998f97e95b" containerID="54db3241cd351152e586e0a4934bc27134069108503f0ad2f6832d8daa165748" exitCode=0 Feb 22 00:33:32 crc kubenswrapper[4857]: I0222 00:33:32.282472 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqlfh" event={"ID":"7cc49fb4-2ca3-49dc-8a95-30998f97e95b","Type":"ContainerDied","Data":"54db3241cd351152e586e0a4934bc27134069108503f0ad2f6832d8daa165748"} Feb 22 00:33:32 crc kubenswrapper[4857]: I0222 00:33:32.305000 4857 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nsd69" Feb 22 00:33:32 crc kubenswrapper[4857]: I0222 00:33:32.572517 4857 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nsd69"] Feb 22 00:33:32 crc kubenswrapper[4857]: W0222 00:33:32.583822 4857 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc082b19_503f_4c5e_af76_99fe7e2bf0fd.slice/crio-c43c76f467cb502fdeef7b4f1ba8e3e751b501a5f463fe7ea61e2b8c97c54c75 WatchSource:0}: Error finding container c43c76f467cb502fdeef7b4f1ba8e3e751b501a5f463fe7ea61e2b8c97c54c75: Status 404 returned error can't find the container with id c43c76f467cb502fdeef7b4f1ba8e3e751b501a5f463fe7ea61e2b8c97c54c75 Feb 22 00:33:33 crc kubenswrapper[4857]: I0222 00:33:33.292706 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqlfh" event={"ID":"7cc49fb4-2ca3-49dc-8a95-30998f97e95b","Type":"ContainerStarted","Data":"2f0d25bda42b0eeccb5c62666f835b4ffe9aeba7e6544ba257abfc145e5e4479"} Feb 22 00:33:33 crc kubenswrapper[4857]: I0222 00:33:33.293816 4857 generic.go:334] "Generic (PLEG): container finished" podID="fc082b19-503f-4c5e-af76-99fe7e2bf0fd" containerID="83cf0d3ec47e7771b467ba3508296fec9c38cc8fff86bf0d9b1b9d11b1f880a2" exitCode=0 Feb 22 00:33:33 crc kubenswrapper[4857]: I0222 00:33:33.293843 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsd69" event={"ID":"fc082b19-503f-4c5e-af76-99fe7e2bf0fd","Type":"ContainerDied","Data":"83cf0d3ec47e7771b467ba3508296fec9c38cc8fff86bf0d9b1b9d11b1f880a2"} Feb 22 00:33:33 crc kubenswrapper[4857]: I0222 00:33:33.293871 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsd69" event={"ID":"fc082b19-503f-4c5e-af76-99fe7e2bf0fd","Type":"ContainerStarted","Data":"c43c76f467cb502fdeef7b4f1ba8e3e751b501a5f463fe7ea61e2b8c97c54c75"} Feb 22 00:33:33 crc kubenswrapper[4857]: I0222 00:33:33.322732 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qqlfh" podStartSLOduration=2.6914064189999998 podStartE2EDuration="5.322716299s" podCreationTimestamp="2026-02-22 00:33:28 +0000 UTC" firstStartedPulling="2026-02-22 00:33:30.270302139 +0000 UTC m=+1617.909031392" lastFinishedPulling="2026-02-22 00:33:32.901612019 +0000 UTC m=+1620.540341272" observedRunningTime="2026-02-22 00:33:33.316420802 +0000 UTC m=+1620.955150055" watchObservedRunningTime="2026-02-22 00:33:33.322716299 +0000 UTC m=+1620.961445552" Feb 22 00:33:34 crc kubenswrapper[4857]: I0222 00:33:34.300093 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsd69" event={"ID":"fc082b19-503f-4c5e-af76-99fe7e2bf0fd","Type":"ContainerStarted","Data":"7204d55fd153b1de3a8c77adabdb5654c9bc7224ea9d202eadd2bb6d37ddf917"} Feb 22 00:33:35 crc kubenswrapper[4857]: I0222 00:33:35.306184 4857 generic.go:334] "Generic (PLEG): container finished" podID="fc082b19-503f-4c5e-af76-99fe7e2bf0fd" containerID="7204d55fd153b1de3a8c77adabdb5654c9bc7224ea9d202eadd2bb6d37ddf917" exitCode=0 Feb 22 00:33:35 crc kubenswrapper[4857]: I0222 00:33:35.306234 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsd69" event={"ID":"fc082b19-503f-4c5e-af76-99fe7e2bf0fd","Type":"ContainerDied","Data":"7204d55fd153b1de3a8c77adabdb5654c9bc7224ea9d202eadd2bb6d37ddf917"} Feb 22 00:33:36 crc kubenswrapper[4857]: I0222 00:33:36.076989 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:33:36 crc kubenswrapper[4857]: E0222 00:33:36.077477 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:33:36 crc kubenswrapper[4857]: E0222 00:33:36.078535 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:33:36 crc kubenswrapper[4857]: I0222 00:33:36.318194 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsd69" event={"ID":"fc082b19-503f-4c5e-af76-99fe7e2bf0fd","Type":"ContainerStarted","Data":"83acb51a888525e6b48256becf83281893fde458e0c952e9f50764661abb2da2"} Feb 22 00:33:36 crc kubenswrapper[4857]: I0222 00:33:36.344490 4857 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nsd69" podStartSLOduration=2.982855816 podStartE2EDuration="5.344474322s" podCreationTimestamp="2026-02-22 00:33:31 +0000 UTC" firstStartedPulling="2026-02-22 00:33:33.298600266 +0000 UTC m=+1620.937329519" lastFinishedPulling="2026-02-22 00:33:35.660218772 +0000 UTC m=+1623.298948025" observedRunningTime="2026-02-22 00:33:36.340497528 +0000 UTC m=+1623.979226791" watchObservedRunningTime="2026-02-22 00:33:36.344474322 +0000 UTC m=+1623.983203575" Feb 22 00:33:39 crc kubenswrapper[4857]: E0222 00:33:39.079887 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:33:39 crc kubenswrapper[4857]: I0222 00:33:39.336667 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qqlfh" Feb 22 00:33:39 crc kubenswrapper[4857]: I0222 00:33:39.336721 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qqlfh" Feb 22 00:33:39 crc kubenswrapper[4857]: I0222 00:33:39.377122 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qqlfh" Feb 22 00:33:40 crc kubenswrapper[4857]: I0222 00:33:40.376445 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qqlfh" Feb 22 00:33:40 crc kubenswrapper[4857]: I0222 00:33:40.577484 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qqlfh"] Feb 22 00:33:42 crc kubenswrapper[4857]: I0222 00:33:42.306225 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nsd69" Feb 22 00:33:42 crc kubenswrapper[4857]: I0222 00:33:42.306633 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nsd69" Feb 22 00:33:42 crc kubenswrapper[4857]: I0222 00:33:42.342962 4857 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nsd69" Feb 22 00:33:42 crc kubenswrapper[4857]: I0222 00:33:42.350443 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qqlfh" podUID="7cc49fb4-2ca3-49dc-8a95-30998f97e95b" containerName="registry-server" containerID="cri-o://2f0d25bda42b0eeccb5c62666f835b4ffe9aeba7e6544ba257abfc145e5e4479" gracePeriod=2 Feb 22 00:33:42 crc kubenswrapper[4857]: I0222 00:33:42.414321 4857 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nsd69" Feb 22 00:33:42 crc kubenswrapper[4857]: I0222 00:33:42.969471 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nsd69"] Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.193368 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qqlfh" Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.332131 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cc49fb4-2ca3-49dc-8a95-30998f97e95b-utilities\") pod \"7cc49fb4-2ca3-49dc-8a95-30998f97e95b\" (UID: \"7cc49fb4-2ca3-49dc-8a95-30998f97e95b\") " Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.332184 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cc49fb4-2ca3-49dc-8a95-30998f97e95b-catalog-content\") pod \"7cc49fb4-2ca3-49dc-8a95-30998f97e95b\" (UID: \"7cc49fb4-2ca3-49dc-8a95-30998f97e95b\") " Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.332208 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tz4lj\" (UniqueName: \"kubernetes.io/projected/7cc49fb4-2ca3-49dc-8a95-30998f97e95b-kube-api-access-tz4lj\") pod \"7cc49fb4-2ca3-49dc-8a95-30998f97e95b\" (UID: \"7cc49fb4-2ca3-49dc-8a95-30998f97e95b\") " Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.332945 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cc49fb4-2ca3-49dc-8a95-30998f97e95b-utilities" (OuterVolumeSpecName: "utilities") pod "7cc49fb4-2ca3-49dc-8a95-30998f97e95b" (UID: "7cc49fb4-2ca3-49dc-8a95-30998f97e95b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.340006 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cc49fb4-2ca3-49dc-8a95-30998f97e95b-kube-api-access-tz4lj" (OuterVolumeSpecName: "kube-api-access-tz4lj") pod "7cc49fb4-2ca3-49dc-8a95-30998f97e95b" (UID: "7cc49fb4-2ca3-49dc-8a95-30998f97e95b"). InnerVolumeSpecName "kube-api-access-tz4lj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.356509 4857 generic.go:334] "Generic (PLEG): container finished" podID="7cc49fb4-2ca3-49dc-8a95-30998f97e95b" containerID="2f0d25bda42b0eeccb5c62666f835b4ffe9aeba7e6544ba257abfc145e5e4479" exitCode=0 Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.356547 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqlfh" event={"ID":"7cc49fb4-2ca3-49dc-8a95-30998f97e95b","Type":"ContainerDied","Data":"2f0d25bda42b0eeccb5c62666f835b4ffe9aeba7e6544ba257abfc145e5e4479"} Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.356576 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qqlfh" Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.356606 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqlfh" event={"ID":"7cc49fb4-2ca3-49dc-8a95-30998f97e95b","Type":"ContainerDied","Data":"52334db83e4b813c31cd32ad5767eccce3dd6399e56a7b47ce1d8116967ec518"} Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.356630 4857 scope.go:117] "RemoveContainer" containerID="2f0d25bda42b0eeccb5c62666f835b4ffe9aeba7e6544ba257abfc145e5e4479" Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.376001 4857 scope.go:117] "RemoveContainer" containerID="54db3241cd351152e586e0a4934bc27134069108503f0ad2f6832d8daa165748" Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.398850 4857 scope.go:117] "RemoveContainer" containerID="aaab1793ec017cde895a0ea311fb16c4a9ddb68484a853e2949effb7b20e9644" Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.428416 4857 scope.go:117] "RemoveContainer" containerID="2f0d25bda42b0eeccb5c62666f835b4ffe9aeba7e6544ba257abfc145e5e4479" Feb 22 00:33:43 crc kubenswrapper[4857]: E0222 00:33:43.428820 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f0d25bda42b0eeccb5c62666f835b4ffe9aeba7e6544ba257abfc145e5e4479\": container with ID starting with 2f0d25bda42b0eeccb5c62666f835b4ffe9aeba7e6544ba257abfc145e5e4479 not found: ID does not exist" containerID="2f0d25bda42b0eeccb5c62666f835b4ffe9aeba7e6544ba257abfc145e5e4479" Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.428852 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f0d25bda42b0eeccb5c62666f835b4ffe9aeba7e6544ba257abfc145e5e4479"} err="failed to get container status \"2f0d25bda42b0eeccb5c62666f835b4ffe9aeba7e6544ba257abfc145e5e4479\": rpc error: code = NotFound desc = could not find container \"2f0d25bda42b0eeccb5c62666f835b4ffe9aeba7e6544ba257abfc145e5e4479\": container with ID starting with 2f0d25bda42b0eeccb5c62666f835b4ffe9aeba7e6544ba257abfc145e5e4479 not found: ID does not exist" Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.428875 4857 scope.go:117] "RemoveContainer" containerID="54db3241cd351152e586e0a4934bc27134069108503f0ad2f6832d8daa165748" Feb 22 00:33:43 crc kubenswrapper[4857]: E0222 00:33:43.429283 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54db3241cd351152e586e0a4934bc27134069108503f0ad2f6832d8daa165748\": container with ID starting with 54db3241cd351152e586e0a4934bc27134069108503f0ad2f6832d8daa165748 not found: ID does not exist" containerID="54db3241cd351152e586e0a4934bc27134069108503f0ad2f6832d8daa165748" Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.429305 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54db3241cd351152e586e0a4934bc27134069108503f0ad2f6832d8daa165748"} err="failed to get container status \"54db3241cd351152e586e0a4934bc27134069108503f0ad2f6832d8daa165748\": rpc error: code = NotFound desc = could not find container \"54db3241cd351152e586e0a4934bc27134069108503f0ad2f6832d8daa165748\": container with ID starting with 54db3241cd351152e586e0a4934bc27134069108503f0ad2f6832d8daa165748 not found: ID does not exist" Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.429320 4857 scope.go:117] "RemoveContainer" containerID="aaab1793ec017cde895a0ea311fb16c4a9ddb68484a853e2949effb7b20e9644" Feb 22 00:33:43 crc kubenswrapper[4857]: E0222 00:33:43.429766 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aaab1793ec017cde895a0ea311fb16c4a9ddb68484a853e2949effb7b20e9644\": container with ID starting with aaab1793ec017cde895a0ea311fb16c4a9ddb68484a853e2949effb7b20e9644 not found: ID does not exist" containerID="aaab1793ec017cde895a0ea311fb16c4a9ddb68484a853e2949effb7b20e9644" Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.429792 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aaab1793ec017cde895a0ea311fb16c4a9ddb68484a853e2949effb7b20e9644"} err="failed to get container status \"aaab1793ec017cde895a0ea311fb16c4a9ddb68484a853e2949effb7b20e9644\": rpc error: code = NotFound desc = could not find container \"aaab1793ec017cde895a0ea311fb16c4a9ddb68484a853e2949effb7b20e9644\": container with ID starting with aaab1793ec017cde895a0ea311fb16c4a9ddb68484a853e2949effb7b20e9644 not found: ID does not exist" Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.433219 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7cc49fb4-2ca3-49dc-8a95-30998f97e95b-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.433237 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tz4lj\" (UniqueName: \"kubernetes.io/projected/7cc49fb4-2ca3-49dc-8a95-30998f97e95b-kube-api-access-tz4lj\") on node \"crc\" DevicePath \"\"" Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.476682 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cc49fb4-2ca3-49dc-8a95-30998f97e95b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7cc49fb4-2ca3-49dc-8a95-30998f97e95b" (UID: "7cc49fb4-2ca3-49dc-8a95-30998f97e95b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.534088 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7cc49fb4-2ca3-49dc-8a95-30998f97e95b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.692488 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qqlfh"] Feb 22 00:33:43 crc kubenswrapper[4857]: I0222 00:33:43.697561 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qqlfh"] Feb 22 00:33:44 crc kubenswrapper[4857]: I0222 00:33:44.365873 4857 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nsd69" podUID="fc082b19-503f-4c5e-af76-99fe7e2bf0fd" containerName="registry-server" containerID="cri-o://83acb51a888525e6b48256becf83281893fde458e0c952e9f50764661abb2da2" gracePeriod=2 Feb 22 00:33:44 crc kubenswrapper[4857]: I0222 00:33:44.710468 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nsd69" Feb 22 00:33:44 crc kubenswrapper[4857]: I0222 00:33:44.850507 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrmz7\" (UniqueName: \"kubernetes.io/projected/fc082b19-503f-4c5e-af76-99fe7e2bf0fd-kube-api-access-lrmz7\") pod \"fc082b19-503f-4c5e-af76-99fe7e2bf0fd\" (UID: \"fc082b19-503f-4c5e-af76-99fe7e2bf0fd\") " Feb 22 00:33:44 crc kubenswrapper[4857]: I0222 00:33:44.850561 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc082b19-503f-4c5e-af76-99fe7e2bf0fd-utilities\") pod \"fc082b19-503f-4c5e-af76-99fe7e2bf0fd\" (UID: \"fc082b19-503f-4c5e-af76-99fe7e2bf0fd\") " Feb 22 00:33:44 crc kubenswrapper[4857]: I0222 00:33:44.850636 4857 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc082b19-503f-4c5e-af76-99fe7e2bf0fd-catalog-content\") pod \"fc082b19-503f-4c5e-af76-99fe7e2bf0fd\" (UID: \"fc082b19-503f-4c5e-af76-99fe7e2bf0fd\") " Feb 22 00:33:44 crc kubenswrapper[4857]: I0222 00:33:44.852244 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc082b19-503f-4c5e-af76-99fe7e2bf0fd-utilities" (OuterVolumeSpecName: "utilities") pod "fc082b19-503f-4c5e-af76-99fe7e2bf0fd" (UID: "fc082b19-503f-4c5e-af76-99fe7e2bf0fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:33:44 crc kubenswrapper[4857]: I0222 00:33:44.857260 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc082b19-503f-4c5e-af76-99fe7e2bf0fd-kube-api-access-lrmz7" (OuterVolumeSpecName: "kube-api-access-lrmz7") pod "fc082b19-503f-4c5e-af76-99fe7e2bf0fd" (UID: "fc082b19-503f-4c5e-af76-99fe7e2bf0fd"). InnerVolumeSpecName "kube-api-access-lrmz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 22 00:33:44 crc kubenswrapper[4857]: I0222 00:33:44.917243 4857 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc082b19-503f-4c5e-af76-99fe7e2bf0fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc082b19-503f-4c5e-af76-99fe7e2bf0fd" (UID: "fc082b19-503f-4c5e-af76-99fe7e2bf0fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 22 00:33:44 crc kubenswrapper[4857]: I0222 00:33:44.951722 4857 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrmz7\" (UniqueName: \"kubernetes.io/projected/fc082b19-503f-4c5e-af76-99fe7e2bf0fd-kube-api-access-lrmz7\") on node \"crc\" DevicePath \"\"" Feb 22 00:33:44 crc kubenswrapper[4857]: I0222 00:33:44.951763 4857 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc082b19-503f-4c5e-af76-99fe7e2bf0fd-utilities\") on node \"crc\" DevicePath \"\"" Feb 22 00:33:44 crc kubenswrapper[4857]: I0222 00:33:44.951773 4857 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc082b19-503f-4c5e-af76-99fe7e2bf0fd-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 22 00:33:45 crc kubenswrapper[4857]: I0222 00:33:45.083769 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cc49fb4-2ca3-49dc-8a95-30998f97e95b" path="/var/lib/kubelet/pods/7cc49fb4-2ca3-49dc-8a95-30998f97e95b/volumes" Feb 22 00:33:45 crc kubenswrapper[4857]: I0222 00:33:45.373655 4857 generic.go:334] "Generic (PLEG): container finished" podID="fc082b19-503f-4c5e-af76-99fe7e2bf0fd" containerID="83acb51a888525e6b48256becf83281893fde458e0c952e9f50764661abb2da2" exitCode=0 Feb 22 00:33:45 crc kubenswrapper[4857]: I0222 00:33:45.373704 4857 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nsd69" Feb 22 00:33:45 crc kubenswrapper[4857]: I0222 00:33:45.373702 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsd69" event={"ID":"fc082b19-503f-4c5e-af76-99fe7e2bf0fd","Type":"ContainerDied","Data":"83acb51a888525e6b48256becf83281893fde458e0c952e9f50764661abb2da2"} Feb 22 00:33:45 crc kubenswrapper[4857]: I0222 00:33:45.373840 4857 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nsd69" event={"ID":"fc082b19-503f-4c5e-af76-99fe7e2bf0fd","Type":"ContainerDied","Data":"c43c76f467cb502fdeef7b4f1ba8e3e751b501a5f463fe7ea61e2b8c97c54c75"} Feb 22 00:33:45 crc kubenswrapper[4857]: I0222 00:33:45.373863 4857 scope.go:117] "RemoveContainer" containerID="83acb51a888525e6b48256becf83281893fde458e0c952e9f50764661abb2da2" Feb 22 00:33:45 crc kubenswrapper[4857]: I0222 00:33:45.393217 4857 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nsd69"] Feb 22 00:33:45 crc kubenswrapper[4857]: I0222 00:33:45.395306 4857 scope.go:117] "RemoveContainer" containerID="7204d55fd153b1de3a8c77adabdb5654c9bc7224ea9d202eadd2bb6d37ddf917" Feb 22 00:33:45 crc kubenswrapper[4857]: I0222 00:33:45.398268 4857 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nsd69"] Feb 22 00:33:45 crc kubenswrapper[4857]: I0222 00:33:45.413312 4857 scope.go:117] "RemoveContainer" containerID="83cf0d3ec47e7771b467ba3508296fec9c38cc8fff86bf0d9b1b9d11b1f880a2" Feb 22 00:33:45 crc kubenswrapper[4857]: I0222 00:33:45.435821 4857 scope.go:117] "RemoveContainer" containerID="83acb51a888525e6b48256becf83281893fde458e0c952e9f50764661abb2da2" Feb 22 00:33:45 crc kubenswrapper[4857]: E0222 00:33:45.436351 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83acb51a888525e6b48256becf83281893fde458e0c952e9f50764661abb2da2\": container with ID starting with 83acb51a888525e6b48256becf83281893fde458e0c952e9f50764661abb2da2 not found: ID does not exist" containerID="83acb51a888525e6b48256becf83281893fde458e0c952e9f50764661abb2da2" Feb 22 00:33:45 crc kubenswrapper[4857]: I0222 00:33:45.436379 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83acb51a888525e6b48256becf83281893fde458e0c952e9f50764661abb2da2"} err="failed to get container status \"83acb51a888525e6b48256becf83281893fde458e0c952e9f50764661abb2da2\": rpc error: code = NotFound desc = could not find container \"83acb51a888525e6b48256becf83281893fde458e0c952e9f50764661abb2da2\": container with ID starting with 83acb51a888525e6b48256becf83281893fde458e0c952e9f50764661abb2da2 not found: ID does not exist" Feb 22 00:33:45 crc kubenswrapper[4857]: I0222 00:33:45.436400 4857 scope.go:117] "RemoveContainer" containerID="7204d55fd153b1de3a8c77adabdb5654c9bc7224ea9d202eadd2bb6d37ddf917" Feb 22 00:33:45 crc kubenswrapper[4857]: E0222 00:33:45.436796 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7204d55fd153b1de3a8c77adabdb5654c9bc7224ea9d202eadd2bb6d37ddf917\": container with ID starting with 7204d55fd153b1de3a8c77adabdb5654c9bc7224ea9d202eadd2bb6d37ddf917 not found: ID does not exist" containerID="7204d55fd153b1de3a8c77adabdb5654c9bc7224ea9d202eadd2bb6d37ddf917" Feb 22 00:33:45 crc kubenswrapper[4857]: I0222 00:33:45.436819 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7204d55fd153b1de3a8c77adabdb5654c9bc7224ea9d202eadd2bb6d37ddf917"} err="failed to get container status \"7204d55fd153b1de3a8c77adabdb5654c9bc7224ea9d202eadd2bb6d37ddf917\": rpc error: code = NotFound desc = could not find container \"7204d55fd153b1de3a8c77adabdb5654c9bc7224ea9d202eadd2bb6d37ddf917\": container with ID starting with 7204d55fd153b1de3a8c77adabdb5654c9bc7224ea9d202eadd2bb6d37ddf917 not found: ID does not exist" Feb 22 00:33:45 crc kubenswrapper[4857]: I0222 00:33:45.436833 4857 scope.go:117] "RemoveContainer" containerID="83cf0d3ec47e7771b467ba3508296fec9c38cc8fff86bf0d9b1b9d11b1f880a2" Feb 22 00:33:45 crc kubenswrapper[4857]: E0222 00:33:45.437090 4857 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83cf0d3ec47e7771b467ba3508296fec9c38cc8fff86bf0d9b1b9d11b1f880a2\": container with ID starting with 83cf0d3ec47e7771b467ba3508296fec9c38cc8fff86bf0d9b1b9d11b1f880a2 not found: ID does not exist" containerID="83cf0d3ec47e7771b467ba3508296fec9c38cc8fff86bf0d9b1b9d11b1f880a2" Feb 22 00:33:45 crc kubenswrapper[4857]: I0222 00:33:45.437108 4857 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83cf0d3ec47e7771b467ba3508296fec9c38cc8fff86bf0d9b1b9d11b1f880a2"} err="failed to get container status \"83cf0d3ec47e7771b467ba3508296fec9c38cc8fff86bf0d9b1b9d11b1f880a2\": rpc error: code = NotFound desc = could not find container \"83cf0d3ec47e7771b467ba3508296fec9c38cc8fff86bf0d9b1b9d11b1f880a2\": container with ID starting with 83cf0d3ec47e7771b467ba3508296fec9c38cc8fff86bf0d9b1b9d11b1f880a2 not found: ID does not exist" Feb 22 00:33:47 crc kubenswrapper[4857]: I0222 00:33:47.084285 4857 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc082b19-503f-4c5e-af76-99fe7e2bf0fd" path="/var/lib/kubelet/pods/fc082b19-503f-4c5e-af76-99fe7e2bf0fd/volumes" Feb 22 00:33:48 crc kubenswrapper[4857]: E0222 00:33:48.078723 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:33:50 crc kubenswrapper[4857]: I0222 00:33:50.077418 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:33:50 crc kubenswrapper[4857]: E0222 00:33:50.077620 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:33:52 crc kubenswrapper[4857]: E0222 00:33:52.077942 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:34:02 crc kubenswrapper[4857]: E0222 00:34:02.079444 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:34:04 crc kubenswrapper[4857]: I0222 00:34:04.077367 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:34:04 crc kubenswrapper[4857]: E0222 00:34:04.077855 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:34:07 crc kubenswrapper[4857]: E0222 00:34:07.079251 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:34:15 crc kubenswrapper[4857]: E0222 00:34:15.079984 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:34:17 crc kubenswrapper[4857]: I0222 00:34:17.078267 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:34:17 crc kubenswrapper[4857]: E0222 00:34:17.079319 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:34:19 crc kubenswrapper[4857]: E0222 00:34:19.080179 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:34:28 crc kubenswrapper[4857]: I0222 00:34:28.077794 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:34:28 crc kubenswrapper[4857]: E0222 00:34:28.080361 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:34:30 crc kubenswrapper[4857]: E0222 00:34:30.080063 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:34:34 crc kubenswrapper[4857]: E0222 00:34:34.079742 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:34:39 crc kubenswrapper[4857]: I0222 00:34:39.080319 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:34:39 crc kubenswrapper[4857]: E0222 00:34:39.081417 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:34:45 crc kubenswrapper[4857]: E0222 00:34:45.079431 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:34:49 crc kubenswrapper[4857]: E0222 00:34:49.110764 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest" Feb 22 00:34:49 crc kubenswrapper[4857]: E0222 00:34:49.111346 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8qqd2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infrawatch-operators-kqm64_service-telemetry(d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76): ErrImagePull: initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" logger="UnhandledError" Feb 22 00:34:49 crc kubenswrapper[4857]: E0222 00:34:49.112532 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:34:50 crc kubenswrapper[4857]: I0222 00:34:50.077303 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:34:50 crc kubenswrapper[4857]: E0222 00:34:50.077794 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:34:58 crc kubenswrapper[4857]: E0222 00:34:58.105219 4857 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" image="image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest" Feb 22 00:34:58 crc kubenswrapper[4857]: E0222 00:34:58.105860 4857 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bqrvk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infrawatch-operators-9lk9c_service-telemetry(57fe7048-f646-407a-a625-0be29af288ff): ErrImagePull: initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown" logger="UnhandledError" Feb 22 00:34:58 crc kubenswrapper[4857]: E0222 00:34:58.107103 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"initializing source docker://image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest: reading manifest latest in image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index: manifest unknown\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:35:00 crc kubenswrapper[4857]: E0222 00:35:00.079165 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:35:03 crc kubenswrapper[4857]: I0222 00:35:03.088283 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:35:03 crc kubenswrapper[4857]: E0222 00:35:03.088519 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:35:10 crc kubenswrapper[4857]: E0222 00:35:10.080108 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:35:11 crc kubenswrapper[4857]: E0222 00:35:11.085901 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:35:14 crc kubenswrapper[4857]: I0222 00:35:14.077920 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:35:14 crc kubenswrapper[4857]: E0222 00:35:14.078834 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:35:21 crc kubenswrapper[4857]: E0222 00:35:21.079854 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:35:22 crc kubenswrapper[4857]: E0222 00:35:22.080982 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:35:28 crc kubenswrapper[4857]: I0222 00:35:28.077504 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:35:28 crc kubenswrapper[4857]: E0222 00:35:28.078270 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:35:33 crc kubenswrapper[4857]: E0222 00:35:33.086906 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-kqm64" podUID="d0cfdf5a-0cb9-4691-b39f-366aaf5e2b76" Feb 22 00:35:34 crc kubenswrapper[4857]: E0222 00:35:34.079214 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" Feb 22 00:35:43 crc kubenswrapper[4857]: I0222 00:35:43.080541 4857 scope.go:117] "RemoveContainer" containerID="1da306935a42fe8ddb7c5d1a2e1de02f1ed8925dabfb94d33b46a850c6ebc2d6" Feb 22 00:35:43 crc kubenswrapper[4857]: E0222 00:35:43.082290 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gw6k5_openshift-machine-config-operator(2ed2fde8-1449-491b-ae21-2fe11ab008ae)\"" pod="openshift-machine-config-operator/machine-config-daemon-gw6k5" podUID="2ed2fde8-1449-491b-ae21-2fe11ab008ae" Feb 22 00:35:45 crc kubenswrapper[4857]: E0222 00:35:45.084315 4857 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"image-registry.openshift-image-registry.svc:5000/service-telemetry/service-telemetry-framework-index:latest\\\"\"" pod="service-telemetry/infrawatch-operators-9lk9c" podUID="57fe7048-f646-407a-a625-0be29af288ff" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515146447553024463 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015146447554017401 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015146443702016513 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015146443703015464 5ustar corecore